I _think_ the assumption / idea was that "toggle" implies that the output is connected to a pull-up resistor and that the pin either floats or is pulled down to ground, causing the signal to toggle. I don't know if/how that works in practice, though. Guenter > drivers/watchdog/gpio_wdt.c | 13 +++++-------- > 1 file changed, 5 insertions(+), 8 deletions(-) > > diff --git a/drivers/watchdog/gpio_wdt.c b/drivers/watchdog/gpio_wdt.c > index 0923201ce874..f7686688e0e2 100644 > --- a/drivers/watchdog/gpio_wdt.c > +++ b/drivers/watchdog/gpio_wdt.c > @@ -108,7 +108,6 @@ static int gpio_wdt_probe(struct platform_device *pdev) > struct device *dev = &pdev->dev; > struct device_node *np = dev->of_node; > struct gpio_wdt_priv *priv; > - enum gpiod_flags gflags; > unsigned int hw_margin; > const char *algo; > int ret; > @@ -122,17 +121,15 @@ static int gpio_wdt_probe(struct platform_device *pdev) > ret = of_property_read_string(np, "hw_algo", &algo); > if (ret) > return ret; > - if (!strcmp(algo, "toggle")) { > + > + if (!strcmp(algo, "toggle")) > priv->hw_algo = HW_ALGO_TOGGLE; > - gflags = GPIOD_IN; > - } else if (!strcmp(algo, "level")) { > + else if (!strcmp(algo, "level")) > priv->hw_algo = HW_ALGO_LEVEL; > - gflags = GPIOD_OUT_LOW; > - } else { > + else > return -EINVAL; > - } > > - priv->gpiod = devm_gpiod_get(dev, NULL, gflags); > + priv->gpiod = devm_gpiod_get(dev, NULL, GPIOD_OUT_LOW); > if (IS_ERR(priv->gpiod)) > return PTR_ERR(priv->gpiod); > ***** Get the best out of Youtube encoding with GPL QFFT Codecs for : Windows,Linux & Android #RockTheHouseGoogle! Advanced FFT & 3D Audio functions for CPU & GPU https://gpuopen.com/true-audio-next/ Multimedia Codec SDK https://gpuopen.com/advanced-media-framework/ (c)Rupert S https://science.n-helix.com *** Decoder CB 2021 Codecs kAudioDecoderName "FFmpegAudioDecoder" kAudioTracks [{"bytes per channel":2,"bytes per frame":4,"channel layout":"STEREO","channels":2,"codec":"aac","codec delay":0,"discard decoder delay":false,"encryption scheme":"Unencrypted","has extra data":false,"profile":"unknown","sample format":"Signed 16-bit","samples per second":48000,"seek preroll":"0us"}] kVideoDecoderName "MojoVideoDecoder" kVideoPlaybackFreezing 0.10006 kVideoPlaybackRoughness 3.048 kVideoTracks [{"alpha mode":"is_opaque","codec":"h264","coded size":"426x240","color space":"{primaries:BT709, transfer:BT709, matrix:BT709, range:LIMITED}","encryption scheme":"Unencrypted","has extra data":false,"hdr metadata":"unset","natural size":"426x240","orientation":"0°","profile":"h264 baseline","visible rect":"0,0 426x240"}] info "Selected FFmpegAudioDecoder for audio decoding, config: codec: mp3, profile: unknown, bytes_per_channel: 2, channel_layout: STEREO, channels: 2, samples_per_second: 44100, sample_format: Signed 16-bit planar, bytes_per_frame: 4, seek_preroll: 0us, codec_delay: 0, has extra data: false, encryption scheme: Unencrypted, discard decoder delay: true" kAudioDecoderName "FFmpegAudioDecoder" kAudioTracks [{"bytes per channel":2,"bytes per frame":4,"channel layout":"STEREO","channels":2,"codec":"mp3","codec delay":0,"discard decoder delay":true,"encryption scheme":"Unencrypted","has extra data":false,"profile":"unknown","sample format":"Signed 16-bit planar","samples per second":44100,"seek preroll":"0us"}] kBitrate 192000 kAudioDecoderName "FFmpegAudioDecoder" kAudioTracks [{"bytes per channel":4,"bytes per frame":8,"channel layout":"STEREO","channels":2,"codec":"opus","codec delay":312,"discard decoder delay":true,"encryption scheme":"Unencrypted","has extra data":true,"profile":"unknown","sample format":"Float 32-bit","samples per second":48000,"seek preroll":"80000us"}] kVideoDecoderName "VpxVideoDecoder" kVideoTracks [{"alpha mode":"is_opaque","codec":"vp9","coded size":"1920x1080","color space":"{primaries:BT709, transfer:BT709, matrix:BT709, range:LIMITED}","encryption scheme":"Unencrypted","has extra data":false,"hdr metadata":"unset","natural size":"1920x1080","orientation":"0°","profile":"vp9 profile0","visible rect":"0,0 1920x1080"}] kAudioDecoderName "FFmpegAudioDecoder" kAudioTracks [{"bytes per channel":2,"bytes per frame":4,"channel layout":"STEREO","channels":2,"codec":"aac","codec delay":0,"discard decoder delay":false,"encryption scheme":"Unencrypted","has extra data":false,"profile":"unknown","sample format":"Signed 16-bit","samples per second":44100,"seek preroll":"0us"}] kVideoDecoderName "MojoVideoDecoder" kVideoTracks [{"alpha mode":"is_opaque","codec":"h264","coded size":"1920x1080","color space":"{primaries:BT709, transfer:BT709, matrix:BT709, range:LIMITED}","encryption scheme":"Unencrypted","has extra data":false,"hdr metadata":"unset","natural size":"1920x1080","orientation":"0°","profile":"h264 main","visible rect":"0,0 1920x1080"}] *** PlayStation 5 and Xbox Series Spatial Audio Comparison | Technalysis Audio 3D Tested : Tempest,ATMOS,DTX,DTS https://www.youtube.com/watch?v=vsC2orqiCwI * Waves & Shape FFT original QFFT Audio device & CPU/GPU : (c)RS The use of an FFT simple unit to output directly: Sound & other content such as a BLENDER or DAC Content : (c)RS FFT Examples : Analogue smoothed audio .. Using a capacitor on the pin output to a micro diode laser (for analogue Fibre) Digital output using: 8 to 128Bit multiple high frequency burst mode.. (Multi Phase step at higher frequency & smooth interpolation) Analogue wave converted to digital in key steps through a DAC at higher frequency & amplitude. For many systems an analogue wave makes sense when high speed crystal digital is too expensive. Multiple frequency overlapped digital signals with a time formula is also possible. The mic works by calculating angle on a drum... Light.. and timing & dispersion... The audio works by QFFT replication of audio function.. The DAC works by quantifying as Analog digital or Metric Matrix.. The CPU/GPU by interpreting the data of logic, Space & timing... We need to calculate Quantum is not the necessary feature; But it is the highlight of our: Data storage cache. Our Temporary RAM Our Data transport.. Of our fusion future. FFT & fast precise wave operations in SiMD Several features included for Audio & Video : Add to Audio & Video drivers & sdk i love you <3 DL In particular I want Bluetooth audio optimized with SiMD,AVX vector instructions & DSP process drivers.. The opportunity presents itself to improve the DAC; In particular of the Video cards & Audio devices & HardDrives & BDBlueRay Player Record & load functions of the fluctuating laser.. More than that FFT is logical and fast; Precise & adaptive; FP & SiMD present these opportunities with correct FFT operations & SDK's. 3D surround optimised the same, In particular with FFT efficient code, As one imagines video is also effected by FFT .. Video colour & representation & wavelet compression & sharpness restoration.. Vivid presentation of audio & video & 3D objects and texture; For example DOT compression & image,Audio presentation... SSD & HD technology presents unique opportunities for magnetic waves and amplitude speculation & presentation. FFT : FMA : SiMD instructions & speed : application examples : Audio, Colour pallet , Rainbows, LUT, Blood corpuscles with audio & vibration interaction, Rain with environmental effects & gravity.. There are many application examples of transforms in action (More and more complex by example) High performance SIMD modular arithmetic for polynomial evaluation FFT Examples : in the SiMD Folder... Evaluation of FFT and polynomial X array algebra .. is here handled to over 50Bits... As we understand it the maths depends on a 64bit value with a 128Bit .. as explained in the article value have to be in identical ranges bit wise, However odd bit depth sizes are non conforming (God i need coffee!) In one example (page 9) Most of the maths is 64Bit & One value 128Bit "We therefore focus in this article on the use of floating-point (FP) FMA (fused multiply-add) instructions for floating-point based modular arithmetic. Since the FMA instruction performs two operations (a â?? b + c) with one single final rounding, it can indeed be used to design a fast error-free transformation of the product of two floating-point numbers" Our latest addition is a quite detailed example for us High performance SIMD modular arithmetic for polynomial evaluation 2020 Pierre Fortin, Ambroise Fleury, François Lemaire, Michael Monagan https://hal.archives-ouvertes.fr/hal-02552673/document Contains multiple algorithm examples & is open about the computer operations in use. Advanced FFT & 3D Audio functions for CPU & GPU https://gpuopen.com/true-audio-next/ Multimedia Codec SDK https://gpuopen.com/advanced-media-framework/ (c)Rupert S https://science.n-helix.com ***** Lets face it, Realtec could well resource the QFFT Audio device & transformer/DAC (c)Rupert S https://science.n-helix.com document work examples : https://eurekalert.org/pub_releases/2021-01/epfd-lpb010621.php "Light-based processors boost machine-learning processing ECOLE POLYTECHNIQUE FÃ?DÃ?RALE DE LAUSANNE Research News IMAGE IMAGE: SCHEMATIC REPRESENTATION OF A PROCESSOR FOR MATRIX MULTIPLICATIONS WHICH RUNS ON LIGHT. view more CREDIT: UNIVERSITY OF OXFORD The exponential growth of data traffic in our digital age poses some real challenges on processing power. And with the advent of machine learning and AI in, for example, self-driving vehicles and speech recognition, the upward trend is set to continue. All this places a heavy burden on the ability of current computer processors to keep up with demand. Now, an international team of scientists has turned to light to tackle the problem. The researchers developed a new approach and architecture that combines processing and data storage onto a single chip by using light-based, or "photonic" processors, which are shown to surpass conventional electronic chips by processing information much more rapidly and in parallel. The scientists developed a hardware accelerator for so-called matrix-vector multiplications, which are the backbone of neural networks (algorithms that simulate the human brain), which themselves are used for machine-learning algorithms. Since different light wavelengths (colors) don't interfere with each other, the researchers could use multiple wavelengths of light for parallel calculations. But to do this, they used another innovative technology, developed at EPFL, a chip-based "frequency comb", as a light source. "Our study is the first to apply frequency combs in the field of artificially neural networks," says Professor Tobias Kippenberg at EPFL, one the study's leads. Professor Kippenberg's research has pioneered the development of frequency combs. "The frequency comb provides a variety of optical wavelengths that are processed independently of one another in the same photonic chip." "Light-based processors for speeding up tasks in the field of machine learning enable complex mathematical tasks to be processed at high speeds and throughputs," says senior co-author Wolfram Pernice at Münster University, one of the professors who led the research. "This is much faster than conventional chips which rely on electronic data transfer, such as graphic cards or specialized hardware like TPU's (Tensor Processing Unit)." After designing and fabricating the photonic chips, the researchers tested them on a neural network that recognizes of hand-written numbers. Inspired by biology, these networks are a concept in the field of machine learning and are used primarily in the processing of image or audio data. "The convolution operation between input data and one or more filters - which can identify edges in an image, for example, are well suited to our matrix architecture," says Johannes Feldmann, now based at the University of Oxford Department of Materials. Nathan Youngblood (Oxford University) adds: "Exploiting wavelength multiplexing permits higher data rates and computing densities, i.e. operations per area of processor, not previously attained." "This work is a real showcase of European collaborative research," says David Wright at the University of Exeter, who leads the EU project FunComp, which funded the work. "Whilst every research group involved is world-leading in their own way, it was bringing all these parts together that made this work truly possible." The study is published in Nature this week, and has far-reaching applications: higher simultaneous (and energy-saving) processing of data in artificial intelligence, larger neural networks for more accurate forecasts and more precise data analysis, large amounts of clinical data for diagnoses, enhancing rapid evaluation of sensor data in self-driving vehicles, and expanding cloud computing infrastructures with more storage space, computing power, and applications software. ### Reference J. Feldmann, N. Youngblood, M. Karpov, H. Gehring, X. Li, M. Stappers, M. Le Gallo, X. Fu, A. Lukashchuk, A.S. Raja, J. Liu, C.D. Wright, A. Sebastian, T.J. Kippenberg, W.H.P. Pernice, H. Bhaskaran. Parallel convolution processing using an integrated photonic tensor core. Nature 07 January 2021. DOI: 10.1038/s41586-020-03070-1" Time Measurement "Let's Play" Station NitroMagika_LightCaster