This disclosure relates to test and measurement systems and methods, and more particularly to estimating a bit error ratio (BER) of data acquired using a test and measurement instrument.
Bit error ratio (BER) testing measures signal integrity in a signal channel based on the quantity or percentage of transmitted bits received incorrectly. The higher number of incorrect bits indicates poor channel quality.
Conventional BER testing normally involves a dedicated piece of BER testing equipment, not a sampling oscilloscope. The device under test (DUT) sends out a known pattern and the equipment checks for errors by comparing the known pattern to the received pattern.
Issues with this method of testing include having to transmit a large number of bits in order to obtain enough data to compute very small BER values. This takes up a lot of time, which in turn increases the costs. One customer estimate requires four minutes to obtain the value on a manufacturing line. The line must test hundreds of thousands of optical transceivers, or other kinds of DUTs. Another challenge is that the behavioral receiver equalizer is hard or costly to implement in the dedicated BER test system.
The embodiments here reduce the amount of time needed to compute the bit error ratio (BER) for devices under test (DUT), including optical and electrical transceivers. The embodiments here provide devices and methods for performing bit error ratio testing using an oscilloscope, instead of specialized equipment, in a much faster manner than conventional testing. The embodiments here are also hardware agnostic, not being limited to a particular type of oscilloscope.
A test and measurement device 14, typically an oscilloscope (or “scope”) but may comprise other test and measurement devices, captures signals from the DUT and generates one or more waveforms from those signals. The signals may result from the DUT sending and receiving signals with a known device 18, in this case an optical transceiver that has known and stable characteristics. The scope 14 acquires the waveforms from the DUT and the module 16, typically part of the machine learning system, transforms the waveform into a composite waveform image for use in the machine learning system 20. The test automation system 10 provides the temperature to the module 16, and the conventional BER test results to the machine learning system 20 for training. In some embodiments, the module 16 may be part of the scope 14; in some embodiments, the module 16 may be separate from the scope and the machine learning system 20.
The composite waveform image may take many forms. For example, the scope may take a PAM4 (Pulse Amplitude Modulated 4-level) signal and produce three diagrams from the acquired waveforms, each having what is referred to as an ‘eye opening.’ These may be overlaid into a single composite eye diagram with a single composite threshold.
This is one example of a composite waveform image produced by module 16. U.S. patent application Ser. No. 17/592,437, “EYE CLASSES SEPARATOR WITH OVERLAY, AND COMPOSITE, AND DYNAMIC EYE-TRIGGER FOR HUMANS AND MACHINE LEARNING,” filed Feb. 3, 2022, describes methods of generating a composite eye waveform image. This application is incorporated herein in its entirety. This is one example of a type of composite waveform image. Other types include short pattern tensor images, cyclic loop tensors, and eye diagram overlays. U.S. patent application Ser. No. 17/345,283, titled “A CYCLIC LOOP IMAGE REPRESENTATION FOR WAVEFORM DATA,” filed Jun. 11, 2021, describes generating cyclic loop images. U.S. Prov. Pat. App. No. 63/191,908, filed May 21, 2021, describes generating short pattern tensor images. These applications are incorporated herein in their entireties.
The image provided to the machine learning system may comprise tensor images placed onto the different color channels of an RGB image input. The system may create a tensor image for the waveform from channel 1 of the scope and place that image on one of the color inputs, such as the red channel. The system would also create a tensor image for the waveform from channel 2 of the scope and place that on the blue channel. Graphs representing the temperature may then be provided on the green channel, as an example.
However, the machine learning system needs a BER value associated with each composite waveform image for training. Each composite waveform image and its associated BER value are provided to the machine learning system as a data sample for one training cycle. This process repeats for as many transceivers as necessary to train the machine learning system sufficiently to meet a desired level of accuracy. Typically, machine learning systems undergo testing and validation during training to ensure that the prediction accuracy meets the needed requirements. One way of defining the accuracy is to make the scatter plots of actual BER value on the horizontal axis plotted against the machine learning “predicted value” on the vertical axis. One could then compute the standard deviation of the difference between the predicted and actual and use that as an indicator of the accuracy of the measurement. These may be made at training time, using a large waveform training set with associated actual BER values. The waveforms can be fed into the training networks to obtain the predicted values.
It is assumed that the manufacturing process will build the DUTs to achieve a BER value within a relatively small range. Therefore, the machine learning system needs sufficient training to recognize BER for the devices over the range seen during the manufacture of the devices. This requires high enough resolution in the composite waveform image that allows observation of the small range.
One approach to make the smaller BER counts visible in the composite waveform image involves some dynamic range compression. By compressing the dynamic range of the image, it makes the smaller counts more visible in the image relative to the larger counts. The dynamic range compression or correction may take the form of a gamma correction or other log function. Gamma correction refers to a process of encoding linear gain values in images to a non-linear relationship, originally stemming for the use of CRT monitors. Using an image with a dynamic range compression improves the training of the machine learning system so that it more readily determines small changes in the BER around the nominal manufacturing value.
Once trained, the machine learning system will have learned to associate particular composite waveform images with BER values provided for those images. This will allow it to associate new composite waveform images during runtime with BER values with high accuracy. The manufacturing system can then use the estimated BER from the machine learning system as the BER value to determine pass or fail of the DUT.
The scope then acquires the waveform and the machine learning system receives it and sends it to the ML BER system, which generates the composite waveform image of whichever type is used in this instance, as discussed above at 16. The composite waveform image is then sent to the machine learning system 20, along with the temperature from the test automation system. The machine learning system then outputs the BER result from the machine learning system and communicates it back to the test automation software with the BER value. This will then cause the test automation system to pass or fail the DUT.
If the DUT fails, the process could also compute the BER in the conventional manner as a double-check of the BER estimation from the machine learning system, but that may slow the process down too much. Alternatively, that component could undergo conventional testing away from the manufacturing system to compare the BER estimation from the machine learning system to the conventional result. If the values become consistently too far apart, the machine learning system may need to undergo another training process.
The test and measurement device 14 that acquires the waveform and/or the module 16 that renders the composite waveform image may have one or more processors to perform those tasks.
The test and measurement device 14 may also include ports 32 that provide connection to the DUT(s), such as probes that connect to channels in the device, as discussed above. The device may include one or more analog-to-digital converters (ADCs) 34 that convert the incoming analog signal into digitized samples. The device will include at least one memory, such as an acquisition memory 36 that stores the digitized samples and the BER values used in training, etc. Memory 38 may be combined with memory 36 and may store the code to be executed by the processor 30, as well as store user settings, etc. The user inputs 44 may include knobs, buttons, and other controls. The display 42 displays the waveforms and resulting measurements to the user. The display may optionally incorporate user controls if the display 42 is a touch screen display. The device then sends the information from the DUT to the machine learning system through communication port 40.
In this manner, a machine learning system can produce BER estimation values for DUTs much faster than conventional testing. One estimate of conventional testing took 4 minutes per DUT on the manufacturing line. Using machine learning, current testing runs around 0.45 seconds per DUT, approximately 0.2% of the time it takes to test using conventional methods.
Generally, acquiring the waveform may take 2-3 seconds. However, the possibility exists that getting the BER estimation from the machine learning system could take longer than the waveform acquisition. If this case were to arise, one could parallelize the process by acquiring a first waveform and sending it to a first processor for generating of the composite waveform image and sending it to the machine learning system, meanwhile acquiring a second waveform and sending it to a second processor while awaiting the first results. Once all processors have received their results, one could average the BER estimations to provide a final BER value. However, in current testing, the machine learning system provides the BER values much faster than even the time to acquire the waveform.
Aspects of the disclosure may operate on a particularly created hardware, on firmware, digital signal processors, or on a specially programmed general purpose computer including a processor operating according to programmed instructions. The terms controller or processor as used herein are intended to include microprocessors, microcomputers, Application Specific Integrated Circuits (ASICs), and dedicated hardware controllers. One or more aspects of the disclosure may be embodied in computer-usable data and computer-executable instructions, such as in one or more program modules, executed by one or more computers (including monitoring modules), or other devices. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types when executed by a processor in a computer or other device. The computer executable instructions may be stored on a non-transitory computer readable medium such as a hard disk, optical disk, removable storage media, solid state memory, Random Access Memory (RAM), etc. As will be appreciated by one of skill in the art, the functionality of the program modules may be combined or distributed as desired in various aspects. In addition, the functionality may be embodied in whole or in part in firmware or hardware equivalents such as integrated circuits, FPGA (field-programmable gate array), and the like. Particular data structures may be used to more effectively implement one or more aspects of the disclosure, and such data structures are contemplated within the scope of computer executable instructions and computer-usable data described herein.
The disclosed aspects may be implemented, in some cases, in hardware, firmware, software, or any combination thereof. The disclosed aspects may also be implemented as instructions carried by or stored on one or more or non-transitory computer-readable media, which may be read and executed by one or more processors. Such instructions may be referred to as a computer program product. Computer-readable media, as discussed herein, means any media that can be accessed by a computing device. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media.
Computer storage media means any medium that can be used to store computer-readable information. By way of example, and not limitation, computer storage media may include RAM, ROM, Electrically Erasable Programmable Read-Only Memory (EEPROM), flash memory or other memory technology, Compact Disc Read Only Memory (CD-ROM), Digital Video Disc (DVD), or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, and any other volatile or nonvolatile, removable or non-removable media implemented in any technology. Computer storage media excludes signals per se and transitory forms of signal transmission.
Communication media means any media that can be used for the communication of computer-readable information. By way of example, and not limitation, communication media may include coaxial cables, fiber-optic cables, air, or any other media suitable for the communication of electrical, optical, Radio Frequency (RF), infrared, acoustic or other types of signals.
Additionally, this written description makes reference to particular features. It is to be understood that the disclosure in this specification includes all possible combinations of those particular features. For example, where a particular feature is disclosed in the context of a particular aspect, that feature can also be used, to the extent possible, in the context of other aspects.
Also, when reference is made in this application to a method having two or more defined steps or operations, the defined steps or operations can be carried out in any order or simultaneously, unless the context excludes those possibilities.
Illustrative examples of the disclosed technologies are provided below. An embodiment of the technologies may include one or more, and any combination of, the examples described below.
Example 1 is a test and measurement system, comprising: a machine learning system; a test and measurement device including a port configured to connect the test and measurement device to a device under test (DUT); and one or more processors, configured to execute code that causes the one or more processors to: acquire a waveform from the DUT; transform the waveform into a composite waveform image; and send the composite waveform image to the machine learning system to obtain a bit error ratio (BER) value for the DUT.
Example 2 is the test and measurement system of Example 1, wherein the one or more processors are distributed among the test and measurement device, a test automation system and the machine learning system.
Example 3 is the test and measurement system of either of Examples 1 and 2, wherein the composite waveform image comprises one of a short pattern tensor, a cyclic loop tensor or an eye diagram overlay.
Example 4 is the test and measurement system of any of Examples 1 through 3, further comprising a known device connected to the DUT, the known device to transmit a known pattern along a transmission path to the DUT and receive a pattern from the DUT along a reception path.
Example 5 is the test and measurement system of Example 4, wherein the port comprises a first channel input of the test and measurement device connected to the transmission path and a second channel input of the test and measurement device connected to the reception path.
Example 6 is the test and measurement system of any of Examples 1 through 5, wherein the one or more processors are further configured to execute code to cause the one or more processors to train the machine learning system, the code causing the one or more processors to: acquire a training waveform associated with a signal for which a BER value has been measured and stored; transform the training waveform into a training composite waveform image; and provide the training composite waveform image and the stored BER value to the machine learning system as a training data sample.
Example 7 is the test and measurement system of Example 6, wherein the processors are further configured to repeat the code to cause the one or more processors to train the machine learning system until a sufficient number of training samples have been obtained.
Example 8 is the test and measurement system of any of Examples 1 to 7, wherein the one or more processors are further configured to execute code to cause the one or more processors to apply a dynamic range compression on pixels in the composite waveform image to cause smaller counts in the composite waveform image to be more visible relative to the larger counts in the composite waveform image.
Example 9 is the test and measurement system of Example 8, wherein the dynamic range compression comprises one of either a gamma correction or a log function.
Example 10 is the test and measurement system of any of Examples 1 to 9, wherein the one or more processors are further configured to execute code to cause the one or more processors to obtain a temperature when the waveform from the DUT is acquired and send the temperature to the machine learning system with the composite waveform image.
Example 11 is a method of determining a bit error ratio for a device under test (DUT), comprising: acquiring one or more waveforms from the DUT; transforming the one or more waveforms into a composite waveform image; sending the composite waveform image to a machine learning system to obtain a bit error ratio (BER) value for the DUT.
Example 12 the method of Example 11, wherein transforming the one or more waveforms into the composite waveform image comprises transforming the one or more waveforms into one of a short pattern tensor, a cyclic loop tensor or an eye diagram overlay.
Example 13 is the method of Example 11 or 12, further comprising: connecting a known device the DUT; using the known device to transmit a known pattern along a transmission path to the DUT; and receiving a pattern from the DUT along a reception path.
Example 14 is the method of Example 13, further comprising: connecting a first channel of a test and measurement device to the transmission path; connecting a second channel of the test and measurement device to the reception path; and comparing waveforms acquired from the first channel and the second channel.
Example 15 is the method of any of Examples 11 through 12, further comprising: connecting a loop back path between a transmitter and a receiver on the DUT; measuring the BER for that DUT; storing the BER for that DUT; acquiring a training waveform from the DUT; transforming the training waveform into a training composite waveform image; and providing the training composite waveform image and the stored BER value to the machine learning system as a training data sample.
Example 16 is the method of Example 15, further comprising repeating the method on additional DUTs until a sufficient number of training data samples have been obtained.
Example 17 is the method of any of Examples 11 to 16, further comprising applying a dynamic range compression on pixels in the composite waveform image to cause smaller counts to be more visible in the composite waveform image relative to larger counts.
Example 18 is the method of Example 17, wherein applying a dynamic range further comprising applying one of a gamma correction or a log function to the values in the training composite waveform image.
Example 19 is the method of any of Examples 11 to 18, wherein acquiring one or more waveforms from the DUT further comprises obtaining a testing temperature.
Example 20 is the method of Example 19, wherein transforming the one or more waveforms into the composite waveform image further comprises encoding a graphical representation of the testing temperature into the composite waveform image
Although specific aspects of the disclosure have been illustrated and described for purposes of illustration, it will be understood that various modifications may be made without departing from the spirit and scope of the disclosure. Accordingly, the disclosure should not be limited except as by the appended claims.
This disclosure claims benefit of U.S. Provisional Application No. 63/189,886, titled “BIT ERROR RATIO ESTIMATION USING MACHINE LEARNING,” filed on May 18, 2021, the disclosure of which is incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5272723 | Kimoto | Dec 1993 | A |
5397981 | Wiggers | Mar 1995 | A |
5594655 | Berchin | Jan 1997 | A |
6807496 | Pickerd | Oct 2004 | B2 |
7181146 | Yorks | Feb 2007 | B1 |
7298463 | French | Nov 2007 | B2 |
8583395 | Dybsetter | Nov 2013 | B2 |
8861578 | Lusted | Oct 2014 | B1 |
9059803 | Detofsky | Jun 2015 | B2 |
9130751 | Zivny | Sep 2015 | B2 |
9337993 | Lugthart | May 2016 | B1 |
9548858 | Cirit | Jan 2017 | B1 |
9699009 | Ainspan | Jul 2017 | B1 |
9709605 | Alley | Jul 2017 | B2 |
10171161 | Cote | Jan 2019 | B1 |
10209276 | Tan | Feb 2019 | B2 |
10236982 | Zhuge | Mar 2019 | B1 |
10270527 | Mentovich | Apr 2019 | B1 |
10396897 | Malave | Aug 2019 | B1 |
10585121 | Absher | Mar 2020 | B2 |
10727973 | Kumar | Jul 2020 | B1 |
10852323 | Schaefer | Dec 2020 | B2 |
10863255 | Zhang | Dec 2020 | B2 |
11005697 | Liston | May 2021 | B2 |
11040169 | Jung | Jun 2021 | B2 |
11095314 | Medard | Aug 2021 | B2 |
11177986 | Ganesan | Nov 2021 | B1 |
11233561 | O'Shea | Jan 2022 | B1 |
11237190 | Rule | Feb 2022 | B2 |
11336378 | Bottoni | May 2022 | B2 |
11388081 | Sommers | Jul 2022 | B1 |
11476967 | Geng | Oct 2022 | B2 |
11646863 | Balan | May 2023 | B2 |
11695601 | Sudhakaran | Jul 2023 | B2 |
20020063553 | Jungerman | May 2002 | A1 |
20030053170 | Levinson | Mar 2003 | A1 |
20030208330 | Pickerd | Nov 2003 | A1 |
20030220753 | Pickerd | Nov 2003 | A1 |
20040032889 | Hidaka | Feb 2004 | A1 |
20040121733 | Peng | Jun 2004 | A1 |
20040131365 | Lee | Jul 2004 | A1 |
20040136422 | Mahowald | Jul 2004 | A1 |
20040165622 | Lu | Aug 2004 | A1 |
20040223544 | Upton | Nov 2004 | A1 |
20040236527 | Felps | Nov 2004 | A1 |
20050222789 | West | Oct 2005 | A1 |
20050246601 | Waschura | Nov 2005 | A1 |
20050249252 | Sanchez | Nov 2005 | A1 |
20060120720 | Hauenschild | Jun 2006 | A1 |
20080126001 | Murray | May 2008 | A1 |
20080159737 | Noble | Jul 2008 | A1 |
20080212979 | Ota | Sep 2008 | A1 |
20090040335 | Ito | Feb 2009 | A1 |
20110085793 | Oomori | Apr 2011 | A1 |
20110161738 | Zhang | Jun 2011 | A1 |
20110286506 | Libby | Nov 2011 | A1 |
20120226727 | Zivny | Sep 2012 | A1 |
20130046805 | Smith | Feb 2013 | A1 |
20140093233 | Gao | Apr 2014 | A1 |
20140163914 | Alley | Jun 2014 | A1 |
20140343883 | Libby | Nov 2014 | A1 |
20150003505 | Lusted | Jan 2015 | A1 |
20150055694 | Juenemann | Feb 2015 | A1 |
20150207574 | Schoen | Jul 2015 | A1 |
20150350042 | Zivny | Dec 2015 | A1 |
20160191168 | Huang | Jun 2016 | A1 |
20160328501 | Chase | Nov 2016 | A1 |
20180006721 | Ishizaka | Jan 2018 | A1 |
20180045761 | Tan | Feb 2018 | A1 |
20180074096 | Absher | Mar 2018 | A1 |
20180204117 | Brevdo | Jul 2018 | A1 |
20180219636 | Gale | Aug 2018 | A1 |
20180356655 | Welch | Dec 2018 | A1 |
20190038387 | Chu | Feb 2019 | A1 |
20190278500 | Lakshmi | Sep 2019 | A1 |
20190332941 | Towal | Oct 2019 | A1 |
20190370158 | Rivoir | Dec 2019 | A1 |
20190370631 | Fais | Dec 2019 | A1 |
20200035665 | Chuang | Jan 2020 | A1 |
20200057824 | Yeh | Feb 2020 | A1 |
20200166546 | O'Brien | May 2020 | A1 |
20200195353 | Ye | Jun 2020 | A1 |
20200229206 | Badic | Jul 2020 | A1 |
20200313999 | Lee | Oct 2020 | A1 |
20200335029 | Gao | Oct 2020 | A1 |
20210041499 | Ghosal | Feb 2021 | A1 |
20210105548 | Ye | Apr 2021 | A1 |
20210111794 | Huang | Apr 2021 | A1 |
20210160109 | Seol | May 2021 | A1 |
20210167864 | Razzell | Jun 2021 | A1 |
20210314081 | Shattil | Oct 2021 | A1 |
20210389373 | Pickerd | Dec 2021 | A1 |
20210390456 | Pickerd | Dec 2021 | A1 |
20220070040 | Namgoong | Mar 2022 | A1 |
20220076715 | Lee | Mar 2022 | A1 |
20220121388 | Woo | Apr 2022 | A1 |
20220182139 | Zhang | Jun 2022 | A1 |
20220199126 | Lee | Jun 2022 | A1 |
20220200712 | Lillie | Jun 2022 | A1 |
20220215865 | Woo | Jul 2022 | A1 |
20220236326 | Schaefer | Jul 2022 | A1 |
20220239371 | Xu | Jul 2022 | A1 |
20220247648 | Pickerd | Aug 2022 | A1 |
20220311513 | Pickerd | Sep 2022 | A1 |
20220311514 | Smith | Sep 2022 | A1 |
20220334180 | Pickerd | Oct 2022 | A1 |
20220373597 | Agoston | Nov 2022 | A1 |
20220373598 | Tan | Nov 2022 | A1 |
20220385374 | Arikawa | Dec 2022 | A1 |
20220390515 | Pickerd | Dec 2022 | A1 |
20220393914 | Tan | Dec 2022 | A1 |
20220407595 | Varughese | Dec 2022 | A1 |
20230050162 | Tan | Feb 2023 | A1 |
20230050303 | Pickerd | Feb 2023 | A1 |
20230052588 | Sudhakaran | Feb 2023 | A1 |
20230088409 | Parsons | Mar 2023 | A1 |
20230098379 | Smith | Mar 2023 | A1 |
20230194599 | Gilabert | Jun 2023 | A1 |
20230228803 | Sun | Jul 2023 | A1 |
20230239182 | Ikeda | Jul 2023 | A1 |
20230299862 | O'Shea | Sep 2023 | A1 |
20230306578 | Pickerd | Sep 2023 | A1 |
Number | Date | Country |
---|---|---|
107342810 | Nov 2019 | CN |
2743710 | Sep 2018 | EP |
3936877 | Jan 2022 | EP |
6560793 | Aug 2019 | JP |
2021092156 | May 2021 | WO |
2022171645 | Aug 2022 | WO |
2022189613 | Sep 2022 | WO |
Entry |
---|
Varughese, Siddarth, et al., Accelerating Assessments of Optical Components Using Machine Learning: TDECQ as Demonstrated Example, Journal of Lightwave Technology, Jan. 1, 2021, pp. 64-72, vol. 39, No. 1, IEEE. |
Varughese, Siddarth, et al., Accelerating TDECQ Assessments using Convolutional Neural Networks, OFC, Mar. 2020, 3 pages, The Optical Society (OSA). |
Watts et al., “Performance of Single-Mode Fiber Links Using Electronic Feed-Forward and Decision Feedback Equalizers”, 2005, IEEE Photonics Techology Letters, vol. 17, No. 10, pp. 2206-2208 (Year: 2005). |
Echeverri-Chacon et al., “Transmitter and Dispersion Eye Closure Quaternary (TDECQ) and Its Sensitivity to Impairments in PAM4 Waveforms”, 2019, Journal of Lightwave Technology, vol. 37, No. 3, pp. 852-860 (Year: 2019). |
Wang et al., “Intelligent Constellation Diagram Analyzer Using Convolutional Neural Network-Based Deep Learning,” Optics Express, Jul. 24, 2017, vol. 25, No. 15. |
Number | Date | Country | |
---|---|---|---|
20220373597 A1 | Nov 2022 | US |
Number | Date | Country | |
---|---|---|---|
63189886 | May 2021 | US |