USRE40394E1 - Teleoperated robotic sorting system - Google Patents

Teleoperated robotic sorting system Download PDF

Info

Publication number
USRE40394E1
USRE40394E1 US10/252,444 US25244497A USRE40394E US RE40394 E1 USRE40394 E1 US RE40394E1 US 25244497 A US25244497 A US 25244497A US RE40394 E USRE40394 E US RE40394E
Authority
US
United States
Prior art keywords
electromagnetic radiation
wavelength range
material objects
incident electromagnetic
mixture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US10/252,444
Inventor
Charles E. Roos
Edward J. Sommer, Jr.
Robert H. Parrish
James R. Russell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NATIONAL RECOVERY TECHNOLOGIES LLC
Original Assignee
National Recovery Technologies Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National Recovery Technologies Inc filed Critical National Recovery Technologies Inc
Priority to US10/252,444 priority Critical patent/USRE40394E1/en
Application granted granted Critical
Publication of USRE40394E1 publication Critical patent/USRE40394E1/en
Assigned to CALTIUS PARTNERS III, LP, AS AGENT reassignment CALTIUS PARTNERS III, LP, AS AGENT SECURITY AGREEMENT Assignors: NATIONAL RECOVERY TECHNOLOGIES, INC.
Assigned to NATIONAL RECOVERY TECHNOLOGIES, LLC reassignment NATIONAL RECOVERY TECHNOLOGIES, LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: NATIONAL RECOVERY TECHNOLOGIES, INC.
Anticipated expiration legal-status Critical
Assigned to NATIONAL RECOVERY TECHNOLOGIES, LLC reassignment NATIONAL RECOVERY TECHNOLOGIES, LLC RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: CALTIUS PARTNERS III, LP
Assigned to TRUE WEST CAPITAL PARTNERS FUND II, LP reassignment TRUE WEST CAPITAL PARTNERS FUND II, LP SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: NATIONAL RECOVERY TECHNOLOGIES, LLC
Assigned to PNC BANK, NATIONAL ASSOCIATION reassignment PNC BANK, NATIONAL ASSOCIATION SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: EMERGING ACQUISITIONS, LLC, NATIONAL RECOVERY TECHNOLOGIES, LLC, NIHOT RECYCLING TECHNOLOGY B.V., ZERO WASTE ENERGY, LLC
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C5/00Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
    • B07C5/36Sorting apparatus characterised by the means used for distribution
    • B07C5/361Processing or control devices therefor, e.g. escort memory
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07BSEPARATING SOLIDS FROM SOLIDS BY SIEVING, SCREENING, SIFTING OR BY USING GAS CURRENTS; SEPARATING BY OTHER DRY METHODS APPLICABLE TO BULK MATERIAL, e.g. LOOSE ARTICLES FIT TO BE HANDLED LIKE BULK MATERIAL
    • B07B13/00Grading or sorting solid materials by dry methods, not otherwise provided for; Sorting articles otherwise than by indirectly controlled devices
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C5/00Sorting according to a characteristic or feature of the articles or material being sorted, e.g. by control effected by devices which detect or measure such characteristic or feature; Sorting by manually actuated devices, e.g. switches
    • B07C5/36Sorting apparatus characterised by the means used for distribution
    • B07C5/363Sorting apparatus characterised by the means used for distribution by means of air
    • B07C5/367Sorting apparatus characterised by the means used for distribution by means of air using a plurality of separation means
    • B07C5/368Sorting apparatus characterised by the means used for distribution by means of air using a plurality of separation means actuated independently
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B07SEPARATING SOLIDS FROM SOLIDS; SORTING
    • B07CPOSTAL SORTING; SORTING INDIVIDUAL ARTICLES, OR BULK MATERIAL FIT TO BE SORTED PIECE-MEAL, e.g. BY PICKING
    • B07C7/00Sorting by hand only e.g. of mail
    • B07C7/005Computer assisted manual sorting, e.g. for mail
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S209/00Classifying, separating, and assorting solids
    • Y10S209/942Operator selects destination of item

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Sorting Of Articles (AREA)

Abstract

A method and apparatus are disclosed for classifying materials utilizing a computerized touch sensitive screen or other computerized pointing device for operator identification and electronic marking of spatial coordinates of materials to be extracted. An operator positioned at a computerized touch sensitive screen views electronic images of the mixture of materials to be sorted as they are conveyed past a sensor array which transmits sequences of images of the mixture either directly or through a computer to the touch sensitive display screen. The operator manually “touches” objects displayed on the screen to be extracted from the mixture thereby registering the spatial coordinates of the objects within the computer. The computer then tracks the registered objects as they are conveyed and directs automated devices including mechanical means such as air jets, robotic arms, or other mechanical diverters to extract the registered objects.

Description

This application is a 371 a PCT/US97/19680, filed Nov. 3, 1997 which is a continuation of provisional application 60/030,183 filed Nov. 4, 1996.
The present invention relates generally to a robotic sorting system, and, more particularly to a robotic sorting system suitable for separating recyclable or waste material. This invention was made with Government support under Contract No. DE-FG02-95ER82037, having an effective date of Sep. 1, 1995, a ward ed by the United States Department Of Energy. The Government has certain rights in the invention.
BACKGROUND OF THE INVENTION
In the recycling of waste secondary materials it is very useful to be able to separate mixtures of materials into usable fractions. Such separations are sometimes performed by fully automated systems which use automated sensors for materials identification with subsequent automated extraction of selected materials from mixtures such as disclosed in U.S. Pat. Nos. 5,260,576 and 5,555,984. In many instances automated materials handling and sorting systems are combined with manual handsorting in a semi-automated process such as that disclosed in U.S. Pat. No. 5,411,147. In many other systems, particularly at smaller recycling facilities, most of the sorting is done by manual handsorting.
In a fully automated sorting system identification of materials to be extracted from a conveyed stream is performed by automated sensors which are specific to identification of certain materials. Advantages to these systems are often high speed and the lack of need for manual labor. A disadvantage to these systems is that automated sensors are generally limited in their ability to identify a wide range of materials and therefore have limited applicability to only selected sorting tasks. On the other hand, in manual handsorting the human visual sensory system is used to make identifications and advantageously is capable of efficiently identifying a wide range of various materials to be sorted. However, also in manual handsorting, human hands are used to make the sorting extractions from the waste stream with disadvantages of relatively low capacity compared to the speed of identifications that a human can perform, and secondly, handsorting requires that humans come into contact with a waste stream which in general is unsanitary and often contains hazardous objects such as broken glass and other objects which can easily puncture or cut.
Present sorting technologies for waste materials generally require the presence of an operator on the sorting floor or, in the case of manual handsorting, at the sorting line manually extracting recyclable materials from the waste stream. This generally requires that the equipment operator or manual handsorters be able-bodied. Therefore this type of work has generally not been available to the physically handicapped since the physical demands of the work would be beyond their capabilities or would place them in a dangerous environment because of their physical condition. This situation has lessened job opportunities for the physically handicapped.
Robotic systems have been applied in industry for a number of years typically to reduce human labor, reduce human presence in hazardous or potentially hazardous situations, and to replace humans in tedious repetitive tasks. In U.S. Pat. No. 5,299,693 a robotic system is described for extracting recyclable materials from a waste stream where the recyclable materials physically have a tag coupled to them which provides a non-visual identifying signal which can be received by a sensor for identification and with a robotic arm subsequently guided into the waste stream to retrieve the tagged item. However, the requirement for physical attachment of a signal generating tag to the items to be sorted limits the usefulness of the process and would require a massive change in practices by the packaging industry to provide such signal markers on packaging materials typically found in the waste stream. U.S. Pat. No. 5,411,147 discloses using robotic arms to extract materials from a waste stream although details of the robotic systems and interactive functions with humans are not provided. U.S. Pat. No. 4,942,538 discloses a teleoperated robotic tracker which is guided by a human operator using a joystick type hand controller with video feedback of robotic arm motion to the operator from a camera mounted on the robotic arm. The use of such a system for recyclables sorting would be awkward and slow since the human operator would need to provide continual guidance to the robotic arm throughout the whole sorting process. There are numerous other robotic systems with interactive human operation disclosed in the prior art. However, nowhere in the prior art have the inventors found descriptions of robotic systems for sorting materials where the identification of selected materials to be extracted from a waste stream or other conveyed stream of materials is provided by a human operator utilizing a computerized pointing device such as a touch screen for electronically registering the spatial coordinates of the selected materials with subsequent fully computerized control of a mechanical or other robotic system to acquire and extract the selected materials from the conveyed stream.
SUMMARY OF THE INVENTION
One of the objectives of the present invention is to alleviate one or more of the problems identified above. Another objective of the present invention is to provide a sorting technology which incorporates the sensing flexibility and sensing speed of a human being combined with high speed high capacity mechanical material extraction systems to provide a highly flexible capacity sorting system. A second objective is to provide a sorting technology which incorporates the sensing flexibility and sensing speed of a human being while insulating the human being from contact with the material stream. A third objective is to provide a sorting technology which can be operated by the physically handicapped. A fourth objective is to provide an automated sorting technology capable of being trained by a human operator in order to become fully automated.
These and other objectives are achieved by providing a method for identifying and sorting material objects from a mixture of material objects in which the mixture of material objects is conveyed through an inspection zone. The mixture of material objects is irradiated with incident electromagnetic radiation in the inspection zone. The electromagnetic radiation emanating from the irradiated material objects is measured and processed to produce electronic images suitable for visual display. Interactive selection of material objects is performed and the selected material objects are then separated from the mixture of material objects by an automated device.
Therefore, the herein disclosed invention overcomes the limitations discussed above in automated sorting systems, manual handsorting, and in robotic sorting systems by providing for rapid human identification and selection of objects to be sorted with subsequent rapid extraction of selected objects by fully computerized mechanical means. The invention further provides for use of the human capability for rapid and efficient identification and selection of materials to be sorted without subjecting the human operator to direct contact with the material stream and without requiring that the operator be present on the sorting floor. The invention further provides for human training of the sorting system so that the sorting system retains the flexibility of the human for identifying selected materials for sorting while being fully automated.
The disclosed invention classifies materials by utilizing a computerized touch screen or other computerized pointing device for operator identification and electronic marking of spatial coordinates of materials to be extracted from a mixture of materials with subsequent computerized position tracking and extraction of the marked materials by computer controlled mechanical means. More specifically, an operator positioned at a computerized touch screen views electronic images of the mixture of materials to be sorted as they are conveyed past a sensor array which transmits a sequence of images of the mixture to a touch screen either directly or through a computer. The operator views the touch screen images and manually “touches” objects displayed on the screen to be extracted from the mixture thereby registering the spatial coordinates of the objects within the computer. The computer then tracks the registered objects as they are further conveyed and directs mechanical means such as air jets, robotic arms, or other mechanical diverters to extract the registered objects from the mixture at an appropriate position downstream from the camera position. High speed communications between the touch screen, computer, and mechanical sorting equipment allows that the touch screen monitor can be located remote from the sorting environment such as in an air conditioned office or even at a remote location such as across town or in another locally altogether. Therefore there is no requirement that the operator be in contact with or even near the material stream.
The computer is also capable of “learning” the properties of those objects being selected for extraction by the operator and capable of taking over the selecting process after a sufficient learning process. At that time the system becomes a fully automated sorting system for extraction of those types of selected objects. At any time the computer can be retrained to select different or additional objects by the operator selecting the different or additional objects through the touch screen or other computerized pointing device.
BRIEF DESCRIPTION OF THE DRAWINGS
The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate presently preferred embodiments of the invention, and, together with the general description given above and the detailed description of the preferred embodiments given below, serve to explain the principles of the invention.
FIG. 1 is a perspective view of the apparatus according to a first preferred embodiment of the invention;
FIG. 2 is a schematic diagram showing material objects diverted by an air blast.
FIG. 3 is a bar graph showing the performance of various pointing devices.
FIG. 4 is a bar graph showing the number of correct captures per second using different pointing devices.
FIG. 5 is a perspective view showing a second embodiment of the apparatus according to the invention.
FIG. 6 is a perspective view showing an embodiment of a robotic arm according to the invention.
FIGS. 7a and 7b are diagrams showing the operation of the robotic arm.
FIGS. 8a and 8b are diagrams showing the operation of another embodiment of the robotic arm.
FIG. 9 is a block diagram showing the software interfaces for the first embodiment of the apparatus of the invention.
FIG. 10 is a block diagram showing the hardware components for the first embodiment.
FIG. 11 is a block diagram showing the software interfaces for the second embodiment of the apparatus of the invention.
FIG. 12 is a block diagram showing the hardware components for the second embodiment.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
A first preferred embodiment of the invention is shown in FIG. 1. A mixture of materials 1 to be sorted are conveyed on conveying surface 2 through an inspection zone 3 which is irradiated with electromagnetic radiation from radiation sources 5, for instance halogen lamps when sensing is performed in the visible light range. Other suitable radiation sources 5 can include, for example, Klystron tube (microwave radiation), UV lamp (ultraviolet radiation), IR lamp (infrared radiation), X-Ray tube (X-Ray radiation) and a Radio-Nuclide source (gamma rays). Conveying surface 2 may be a belt conveyor, a slide, a vibrating pan conveyor, a free fall trajectory, or any other means for conveying materials. The material objects comprising the mixture of materials 1 may be conveyed singly or in plurality.
A sensor array 4 (for example, Sony Series 9000 CCD video camera when radiation from sources 5 is in the visible light range) is positioned to view the inspection zone so to provide data arrays corresponding to measurements of electromagnetic radiation emanating from inspection zone 3 and from any materials 1 being conveyed through the inspection zone. The electromagnetic radiation emanating from materials 1 may be reflected radiation, radiation transmitted through materials 1, radiation emitted from materials 1 through fluorescence, or any other forms of radiation resulting from interaction of materials 1 with the incident radiation from sources 5. Sensor array 4 is shown positioned above conveying surface 2 although in practice it may be located at any position required to give the desired view. Sensor array 4 is selected to be sensitive in the wavelength range of electromagnetic radiation emanating from material objects within the inspection zone 3 when irradiated by sources 5 which may be a single source or multiple sources. The geometry of sensor array 4 is determined by the application. For instance the sensor array 4 may be a linear array of sensors or an area array of sensors. It may physically span the full width and/or length of the inspection zone 3 or it may be more compact and use optics to scan the width and/or length of the inspection zone such as with a CCD camera. Sensor array 4 may be positioned on the same side of the inspection zone 3 as is irradiation sources 5 or it may be positioned on the opposite side of inspection zone 3 from sources 5 or positioned at any other location with respect to irradiation sources 5 and inspection zone 3. The effective wavelength ranges of sensor array 4 and sources 5 may each be in any one of the microwave wavelength range, the ultraviolet wavelength range, the visible light wavelength range, the infrared wavelength range, the x-ray wavelength range, or the gamma ray wavelength range or any combination thereof. Sensor array 4 may be fitted with special filters which allow only certain wavelengths of electromagnetic radiation to reach sensor array 4 for measurement.
Sensor array 4 data corresponding to electromagnetic radiation measurements emanating from the inspection zone 3 and from material objects 1 within the inspection zone are transmitted from sensor array 4 to computer 7 and/or touch sensitive screen 9 over transmitting cables 6 or by wireless means. Control of sensor array 4 operation may also be provided by computer 7 over transmitting cables 6 or by wireless transmission. Sensor array 4 data received by computer 7 are processed for analog to digital conversion if not already digital by nature of sensor array 4 and microcomputer processed into digitized electronic images which are transmitted over cables 8 to touch sensitive screen 9 which is capable of electronically registering the coordinates on the screen of a manual touch by a human operator 26. Touch screen 9 displays the digitized images corresponding to the sensor data from sensor array 4 of the inspection zone 3 and the materials to be sorted within the inspection zone 3. Alternatively, the sensor array 4 can transmit the images directly to the touch sensitive screen 9.
Human operator 26 views the electronic images on touch screen 9 and manually touches the image of any material object, in this case object 16, which the operator wishes to be removed from the stream of materials 1. Spatial coordinates describing the location on the touch sensitive screen 9 of the touch by operator 26 are registered by touch sensitive screen 9 and transmitted over cables 8 to computer 7. Computer 7 associates the touch screen coordinates of the registered touch with corresponding spatial location coordinates on conveying surface 2 within inspection zone 3 and further associates any object on conveying surface 2 at that location (in this case object 16) with the touch. Computer 7 then electronically tracks the motion of selected object 16 as it is further conveyed along conveying surface 2 utilizing signals transmitted over cables 28 indicating conveyor speed generated by conveyor speed encoder 27. Similarly, the motion could be tracked by programming into computer 7 a predetermined speed for conveying surface 2 in which case conveyor speed encoder 27 would not be needed. As object 16 leaves the inspection zone 3 and drops off the discharge end of conveying surface 2 it is diverted from its falling trajectory by air blast 15 as shown in FIG. 2. Air blast 15 is generated by computer 7 by sending at the right time control signals over cables 10 (FIG. 1) to the appropriate air solenoid valve 11a within air solenoid array 11 to open air solenoid valve 11a for an appropriate length of time to eject object 16. At such time compressed air from air reservoir 12 flows through air solenoid 11a and is emitted as an air blast from air nozzle 13a within air nozzle array 13, and ejects selected object 16 so that it falls over splitter plate 14 and is segregated from non-ejected material objects 17. In this way selected objects may be sorted away from a mixture of objects.
Further details of the present invention are described in the applicants' report to the U.S. Department of Energy, Phase II application under Grant application number 35853-95-I, titled “Use of Computer Robotics to Reduce Human Contact with the Waste Stream and Lower the Costs for Recyclable Materials,” (hereafter “DOE Report”) and in provisional application 60/030,183 filed Nov. 4, 1996, the disclosures of which are incorporated herein, in their entireties.
Computer 7 may contain a pre-compiled pattern database or identification and pattern recognition algorithms which can perform learning of selections by operator 26 as the operator makes the selections. Such identification and pattern recognition algorithms may be accomplished by computerized neural networks or other such pattern recognition computer code. Identification by pattern recognition of the objects can be performed by using, for example, the edge enhancement and image contour extraction techniques disclosed in the DOE Report at pages 22-29. Further details of pattern recognition and its interaction with robotic systems is described in the published text titled “Robot Vision,” Berthold Klaus Paul Horn, MIT Press, Cambridge, Mass. (1991), the disclosure of which is incorporated herein, in its entirety.
Learning the recognized object patterns can be performed using known neural network or other pattern recognition and learning systems. Neural network techniques for identifying and learning patterns are described in, for example, the published text “Neural Networks for Pattern Recognition,” Christopher M. Bishop, Oxford University Press, New York (1995), (hereafter “Bishop”), the disclosure of which is incorporated herein, in its entirety. Bishop chapters 3-6 describe neural network techniques including single and multiple layer perception as well as different error functions that can be used for training neural networks. Bishop, chapters 9 and 10 describe techniques for learning and generalization in a neural network.
In this case operator 26 will initially make selections of items to be extracted from the mixture of materials such as object 16. As operator 26 makes selections the associated electronic images will be processed through the computer algorithms with the imaging patterns distinctive to the selected items noted by the algorithms. As similar items are repetitively selected by operator 26 the computer algorithms associate the distinctive properties of the imaging patterns with objects to be selected for extraction and begin to electronically select similar patterns for extraction without input from the human operator 26. In this way the computerized system learns those objects to be extracted and after sufficient learning experience can begin sorting without input from operator 26.
The choice of using a touch screen 9 for making the selection of objects to be extracted from the mixture is a matter of preference. Similar pointing devices interfaced to a display screen could be used as a computer mouse, a track ball, a joystick, a touch pad, a light pen, or other such device. The inventors have chosen the touch screen as the preferred pointing device based upon their intensive studies of some of these various types of devices for sorting applications. FIG. 3 shows a bar graph of the results of some of these studies. The Performance Index assesses the overall performance of a human operator using such devices for “pointing and clicking” on certain objects dynamically displayed on a computer monitor in order to select these certain objects for sorting a simulation. The Performance Index considers selection accuracy, selection speed, and ease of use. Four different cases are presented for each pointing device representing various numbers of objects displayed on the monitor screen at a time conveyed at different speeds. From the data it is seen that the touch screen outperformed all other pointing devices by a wide margin.
The inventors have found that a human operator can comfortably view computerized images of a mixture of materials conveyed past a camera and identify and select items to be sorted out of the mixture at a rates up to 2.5 selections per second using a computerized touch screen to make the identifications and selections (see FIG. 4). This is two to five times as fast as industry established typical manual handsorting rates of one item every one to two seconds.
FIG. 9 shows a block diagram of software interfaces for implementing the preferred embodiment of the invention depicted in FIGS. 1 and 2. In FIG. 9 the sensor array 4 sends data through AID converter 46 to microprocessor 47. Human operator 26 (FIG. 1) touches 48 the touch sensitive screen 9 on the displayed image to select object 16 (FIG. 1). This input to touch sensitive screen 9 is tagged 49 by microprocessor 47 as a selection icon 51 which is scrolled 50 in microprocessor 47 host memory and on touch screen 9 as corresponding selected object 16 moves through inspection zone 3 (FIG. 1). Microprocessor 47 also enters a representation of selection icon 51 into the eject queue 52, assigns to the representation of selection icon 51 an appropriate value of air delay 53 to establish proper timing for ejection of selected object 16 (FIG. 1) and an appropriate value of air on time 54 to effect ejection of selected object 16 as it passes in front of ejectors 55 (shown as 11a and 13a in FIG. 2).
FIG. 10 shows a block diagram of a hardware implementation for the preferred embodiment of FIGS. 1, 2, and 9. In this case video camera 56 is used as sensor array 4 (FIGS. 1 and 9) and interfaces to computer 7 (which incorporates microprocessor 47) through a frame grabber card 57. Images from video camera 56 are displayed as digitized or analog electronic images on the touch screen 9. User touch input 48 (FIG. 9) is sent back to computer 7 and processed in conjunction with the electronic images to eject selected object 16 from a mixture of materials 1 (FIG. 1) by activating pneumatic ejectors 55 (shown as 11a and 13a in FIG. 2) through ejector controller card 58 at the appropriate time. An appropriate CCD video camera for this use is a Sony Model DCX-9000 3CCD color progressive scan camera. A representative video card for interfacing with the Sony camera is Coreco Model Ultra II with RGB Digitization Module. A representative computer is Carlo Gavazzi Model 690 Industrial Computer. An appropriate touch screen is PCS Computers Inc. Series G Flat Panel Monitor. An appropriate ejector controller card is National Instruments Model PCI-DIO-96. Appropriate pneumatic ejectors are comprised of MAC Valves Inc. 6200 air solenoid valves coupled to Industrial Spray Products VeeJet Standard nozzles.
FIG. 5 shows a second preferred embodiment of the apparatus according to the invention. A mixture of materials 1 to be sorted are conveyed on conveyor 2 through an inspection zone 3 which is irradiated with electromagnetic radiation from radiation sources 5 (for instance halogen lamps when sensing is performed in the visible light range). Conveyor 2 may be a belt conveyor, a slide, a vibrating pan conveyor, a free fall trajectory, or any other means for conveying materials. The material objects comprising the mixture of materials 1 may be conveyed singly or in plurality.
A sensor array 4 (for example, a Sony Series 9000 CCD video camera when radiation from sources 5 is in the visible light range) is positioned to view the inspection zone 3 so to provide data arrays corresponding to measurements of electromagnetic radiation emanating from inspection zone 3 and from any materials 1 being conveyed through the inspection zone. The electromagnetic radiation emanating from materials 1 may be reflected radiation, radiation transmitted through materials 1, radiation emitted from materials 1 through fluorescence, or any other forms of radiation resulting from interaction of materials 1 with the incident radiation from sources 5. Sensor array 4 is shown positioned above conveyor 2 although in practice it may be located at any position required to give the desired view. Sensor array 4 is selected to be sensitive in the wavelength range of electromagnetic radiation emanating from material objects within the inspection zone 3 when irradiated by sources 5 which may be a single source or multiple sources. The geometry of sensor array 4 is determined by the application. For instance the sensor array 4 may be a linear array of sensors or an area array of sensors. It may physically span the full width and/or length of the inspection zone 3 or it may be more compact and use optics to scan the width and/or length of the inspection zone such as with a CCD camera. Sensor array 4 may be positioned on the same side of the inspection zone 3 as is irradiation sources 5 or it may be positioned on the opposite side of inspection zone 3 from sources 5 or positioned at any other location with respect to irradiation sources 5 and inspection zone 3. The effective wavelength ranges of sensor array 4 and sources 5 may each be in any one of the microwave wavelength range, the ultraviolet wavelength range, visible light wavelength range, the infrared wavelength range, the x-ray wavelength range, or the gamma ray wavelength range or any combination thereof. Sensor array 4 may be fitted with special filters which allow only certain wavelengths of electromagnetic radiation to reach sensor array 4 for measurement.
Sensor array 4 data corresponding to electromagnetic radiation measurements emanating from the inspection zone 3 and from material objects 1 within the inspection zone are transmitted from sensor array 4 to computer 7 over transmitting cables 6 or by wireless transmission. Control of sensor array 4 operation may also be provided by computer 7 over transmitting cables 6. Sensor array 4 data received by computer 7 are processed for analog to digital conversion if not already digital by nature of sensor array 4 and microcomputer processed into electronic images which are transmitted over cables 8 to touch sensitive screen 9 which is capable of electronically registering the coordinates on the screen of a manual touch by a human operator 26. Touch screen 9 displays the images corresponding to the sensor data from sensor array 4 of the inspection zone 3 and the materials to be sorted within the inspection zone 3.
Human operator 26 views the images on touch screen 9 and manually touches the images of any material object, in this case object 25, which the operator wishes to be removed from the stream of materials 1. Spatial coordinates describing the location on the touch screen 9 of the touch by operator 26 are registered by touch screen 9 and transmitted over cables 8 to computer 7. Computer 7 associates the touch screen coordinates of the registered touch with corresponding spatial location coordinates on the surface of conveyor 2 within inspection zone 3 and further associates any object on conveyor 2 at that location (in this case object 25) with the touch. Computer 7 then electronically tracks the motion of selected object 25 as it is further conveyed along conveyor 2 utilizing signals transmitted over cables 28 indicating conveyor speed generated by conveyor speed encoder 27. Similarly, the motion could be tracked by programming into computer 7 a predetermined speed for conveyor 2 in which case conveyor speed encoder 27 would not be needed.
As the selected object 25 is conveyed into the vicinity of robotic arms 18 the computer 7 performs scheduling algorithms to determine which of the robotic arms 18 will be used to most efficiently extract the object 25 from the waste stream. Such scheduling algorithms are specific to the equipment used and materials to be sorted and can be suitably devised by one skilled in the art using, for example, the techniques discussed in “Schedule Efficiency in a Robotic Cell” by Irina Ioachim and Francois Soumis, International Journal of Flexible Manufacturing Systems, Vol. 7, No. 1 (March 1995). The entire contents of this publication is incorporated herein by reference. There may be only one robotic arm 18 or as many robotic arms 18 as required to efficiently sort materials as selected by operator 26 on touch screen 9. For instance if the average acquisition and retrieval time of a robotic arm 18 for the materials to be sorted 1 is two seconds and the operator 26 can make an average of 2.5 selections per second then a minimum of five robotic arms 18 would be required to keep up with operator 26 selections.
As the selected object 25 approaches, the robotic arm 18 chosen by the computer to make the extraction is commanded by computer 7 over cables 10 to rotate (shown as motion 24) into proper position for most efficient timing to make an interception of object 25. Vertical actuator 19 is positioned by computer 7 through control signals over cables 10 to intercept object 25 as it passes under robotic arm 18. Vertical actuator 19 is placed by computer 7 in retracted (lifted) position so the object 25 can pass under it. As object 25 passes under actuator 19 computer 7 signals the actuator via cables 10 to extend downward until end effector 20 (or 21) contacts object 25. Upon contact with object 25 end effector 20 (or 21) acquires object 25 at which time actuator 19 retracts (lifts) and extracts object 25 from the mixture of objects 1. When object 25 is sufficiently lifted so that its lower extent is higher than the top of the other objects 1 the computer 7 signals actuator 19 to translate along robotic arm 18 in one of the directions toward the edge of the conveyor belt 2 as indicted by directional arrows 22. Positioned along each side of conveyor 2 are receiver chutes 29 and 30. Actuator 19 holding object 25 via its end effector 20 (or 21) continues moving until it is above either chute 29 or chute 30 at which time end effector 20 (or 21) releases object 25 into chute 20 or chute 30 which segregates object 25 from the other objects 1 on the conveyor 2 therefore having sorted object 25 from the other objects 1. Actuator 19 then translates back to a position over the conveyor 2 and awaits commands from computer 7 to make another object acquisition and retrieval in response to touches by operator 26 on the touch screen 9.
Computer 7 may contain identification and pattern recognition algorithms which can perform learning of selections by operator 26 as the operator makes the selections. Such identification and pattern recognition algorithms may be accomplished by computerized neural networks or other such pattern recognition computer code, as discussed earlier herein. In this case operator 26 will initially make selections of items to be extracted from the mixture of materials such as object 25. As operator 26 makes selections the associated electronic images will be processed through the computer algorithms with the imaging patterns distinctive to the selected items noted by the algorithms. As similar items are repetitively selected by operator 26 the computer algorithms associate the distinctive properties of the imaging patterns with objects to be selected for extraction and begin to electronically select similar patterns for extraction without input from the human operator 26. In this way the computerized system learns those objects to be extracted and after sufficient learning experience can begin sorting without input from operator 26.
FIGS. 6, 7a, and 7b show views of a robotic arm 18, a vertical actuator 19, and an end effector 20 (or 21) which is a part of the preferred embodiment designed to pick up an object such as object 25 using a suction cup embodiment of the end effector 20. The vertical actuator 19 and end effector 20 (or 21) is also designed to eliminate the need for vertical actuator 19 to translate along robotic arm 18 after acquisition of object 25 in order to discharge object 25 into receiving chute 29 or receiving chute 30 (FIG. 5). The workings of this system is shown in FIGS. 7a and 7b. When robotic arm 18 and vertical actuator 19 are positioned over the object 25 to be picked up, an air cylinder 32 is extended until the suction cup end effector 20 (or 21) contacts object 25. Vacuum is applied to the suction cup by vacuum hose 33 through vacuum pump 31 and through 3-way air valve 38. Air valve 38 at this position is set to allow the vacuum from vacuum pump 31 to pass through it to suction cup end effector 20 (or 21) and to not allow any compressed air from hose 37 to pass. Air cylinder 32 is then retracted so to lift object 25. As air cylinder 32 continues to retract a fixed lever 34 strikes a fixed extension rod 35 which causes the lower portion of the actuator 19 to rotate approximately 90 degrees as shown around a pivot joint 36 so that the suction cup end effector 20 (or 21) points outward toward the receiving chute 29 or the receiving chute 30 (FIG. 5). At this time the air valve 38 is actuated to close off vacuum to the suction cup end effector 20 (or 21) and at the same time apply compressed air from the hose 37 to the suction cup. The compressed air blast 39 forcefully ejects object 25 from the suction cup end effector 20 (or 21) in a direction 40 and into the receiving chute 29 or the receiving chute 30. Computer 7 then positions vertical actuator 19 for retrieval of the next object. This sequence of actions allows the system to pick up object 25 and eject it into receiving chutes 29 or 30 without requiring any translation along robotic arm 18 to the receiving chutes. This allows a significant savings in cycle time and can reduce the number of robotic arms and vertical actuators needed for the system and thereby reducing cost. An appropriate robotic arm is comprised of Hauser HLE 100A Belt Driven Positioner coupled to Origa Rodless Cylinder Model 40-J2220/20x32.75-B-M fitted with suction cup end effector PIAB L5 Vacuum Pump with B20 Suction Cups.
FIGS. 8a and 8b show another embodiment of a vertical actuator 19 and an end effector 20 (or 21) which is designed to retrieve paper sheets or cardboard sheets 44. In this embodiment end effector 20 (or 21) is comprised of heavy duty spikes 43 which can be driven or extracted by air cylinders 42. Air for control of air cylinders 42 is supplied through air hoses 41. In operation the vertical actuator 19 is extended downward until the end effector 20 (or 21) contacts the paper sheet or cardboard 44 at which time the air cylinders 42 are actuated to rapidly and forcefully extend spikes 43 so that they penetrate through sheet 44 as shown. Vertical actuator 19 is then retracted (lifted) to lift sheet 44. Sheet 44 cannot slide off spikes 43 by virtue of their angled geometry. Actuator 19 is then translated to receiving chute 29 or 30 as previously described. When actuator 19 with sheet 44 are positioned over the receiving chute the spikes 43 are retracted by action of the air cylinders 42 and thereby withdrawing from sheet 44. This frees sheet 44 so that it can fall into the receiving chute. Vertical actuator 19 is then repositioned by computer 7 over conveyor 2 to await the next paper sheet or cardboard sheet to be retrieved. End effectors similar in design but much lighter in construction and which are designed for textile handling can be acquired from Techno Sommer, which has an Automatic SCH20 Needle Gripper product.
FIG. 11 shows a block diagram of software interfaces for implementing the preferred embodiment of the invention depicted in FIG. 5. In FIG. 11 the sensor array 4 sends data through A/D converter 46 to microprocessor 47. Human operator 26 (FIG. 5) touches 48 the touch screen 9 on the displayed image to select object 25 (FIG. 5). This input to touch screen 9 is tagged 49 by microprocessor 47 as a selection icon 51 which is scrolled 50 in microprocessor 47 host memory and on touch screen 9 as corresponding selected object 25 moves through inspection zone 3 (FIG. 5). Microprocessor 47 also enters a representation of selection icon 51 into the pick queue 59, assigns to the representation of selection icon 51 appropriate values of position coordinates 60 to establish proper timing for robotic acquisition (picking) of selected object 25 (FIG. 5) and a scheduling assignment 61 to an appropriate robotic arm 18 (FIG. 5) to effect acquisition of selected object 25 as it passes under the appropriate robotic arm 18 and deliver selected object 25 to an appropriate receiving chute 29 or 30 (FIG. 5). Appendix F of the DOE Report discloses sample prototype code for implementing an embodiment of the tagging 49 of a material object by the microprocessor 47 as a selection icon 51 and the subsequent tracking and generation of an ejection signal for the tagged item.
FIG. 12 shows a block diagram of hardware implementation for the preferred embodiment of FIGS. 5 and 11. In this case video camera 56 is used as sensor array 4 (FIGS. 5 and 11) and interfaces to computer 7 (which incorporates microprocessor 47) through a frame grabber card 57. Images from video camera 56 are displayed as digitized or analog electronic images on touch screen 9 (FIG. 11). User touch input 48 (FIG. 11) is sent back to computer 7 and processed in conjunction with the electronic images to extract selected object 25 from a mixture of materials 1 (FIG. 5) by activating robotic arms 62 at the appropriate time through motion controller card(s) 63. An appropriate CCD video camera for this use is a Sony Model DCX-9000 3CCD color progressive scan camera. A representative video card for interfacing with the Sony camera is Coreco Model Ultra II with RGB Digitization Module. A representative computer is Carlo Gavazzi Model 690 Industrial Computer. An appropriate touch screen is PCS Computers Inc. Series G Flat Panel Monitor. An appropriate motion controller card is Parker Compumotor Motion Controller 806450. An appropriate robotic arm is comprised of Hauser HLE 100A Belt Driven Positioner coupled to Origa Rodless Cylinder Model 40-J2220/20x32.75-B-M fitted with suction cup end effector PIAB L5 Vacuum Pump with B20 Suction Cups.
Other embodiments of the invention will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. It is intended that the specification be considered as exemplary only, with the true scope and spirit of the invention being indicated by the following claims.

Claims (59)

1. A method for identifying and sorting selected material objects from a mixture of material objects comprising steps of:
(a) conveying a mixture of material objects into and through an inspection zone;
(b) irradiating said mixture of material objects with incident electromagnetic radiation while in said inspection zone;
(c) measuring the electromagnetic radiation emanating from said irradiated mixture of material objects;
(d) processing said measured measurements of electromagnetic radiation to produce electronic images of said irradiated material objects and presenting said electronic images for visual display;
(e) interactively selecting designating, from said visual display of material objects, selected material objects to be sorted from said mixture of material objects by using a computerized pointing device operable by a human operator; and
(f) using an automated device to separate separating said selected material objects from said mixture of material objects.
2. The method according to claim 1 wherein said interactive selecting of material objects to be sorted is performed by a human operator using a computerized pointing device separating comprises using an automated device to separate said selected material objects from said mixture of material objects.
3. The method according to claim 2 wherein said interactive selecting designating of material objects to be sorted by a human operator is analyzed by computerized learning and control algorithms so that identifications and selections designations for the said selecting designating of material items to be sorted is passed from the human operator to the computerized algorithms.
4. The method according to claim 3 wherein said mixture of material objects is comprised of solid waste materials.
5. The method according to claim 3 wherein said incident electromagnetic radiation is in the microwave wavelength range.
6. The method according to claim 3 wherein said incident electromagnetic radiation is in the ultraviolet wavelength range.
7. The method according to claim 3 wherein said incident electromagnetic radiation is in the visible light wavelength range.
8. The method according to claim 3 wherein said incident electromagnetic radiation is in the infrared wavelength range.
9. The method according to claim 3 wherein said incident electromagnetic radiation is in the x-ray wavelength range.
10. The method according to claim 3 wherein said incident electromagnetic radiation is in the gamma ray wavelength range.
11. The method according to claim 2 wherein said mixture of material objects is comprised of solid waste materials.
12. The method according to claim 2 wherein said incident electromagnetic radiation is in the microwave wavelength range.
13. The method according to claim 2 wherein said incident electromagnetic radiation is in the ultraviolet wavelength range.
14. The method according to claim 2 wherein said incident electromagnetic radiation is in the visible light wavelength range.
15. The method according to claim 2 wherein said incident electromagnetic radiation is in the infrared wavelength range.
16. The method according to claim 2 wherein said incident electromagnetic radiation is in the x-ray wavelength range.
17. The method according to claim 2 wherein said incident electromagnetic radiation is in the gamma ray wavelength range.
18. The method according to claim 1 wherein said interactive selecting designating of material objects to be sorted is performed by a human operator and is analyzed by computerized learning and control algorithms so that decisions identifications and designations for the selecting designating of material items to be sorted is passed from the human operator to the computerized algorithms.
19. The method according to claim 18 wherein said mixture of material objects is comprised of solid waste materials.
20. The method according to claim 18 wherein said incident electromagnetic radiation is in the microwave wavelength range.
21. The method according to claim 18 wherein said incident electromagnetic radiation is in the ultraviolet wavelength range.
22. The method according to claim 18 wherein said incident electromagnetic radiation is in the visible light wavelength range.
23. The method according to claim 18 wherein said incident electromagnetic radiation is in the infrared wavelength range.
24. The method according to claim 18 wherein said incident electromagnetic radiation is in the x-ray wavelength range.
25. The method according to claim 18 wherein said incident electromagnetic radiation is in the gamma ray wavelength range.
26. The method according to claim 1 wherein said mixture of material objects is comprised of solid waste materials.
27. The method according to claim 1 wherein said incident electromagnetic radiation is in the microwave wavelength range.
28. The method according to claim 1 wherein said incident electromagnetic radiation is in the ultraviolet wavelength range.
29. The method according to claim 1 wherein said incident electromagnetic radiation is in the visible light wavelength range.
30. The method according to claim 1 wherein said incident electromagnetic radiation is in the infrared wavelength range.
31. The method according to claim 1 wherein said incident electromagnetic radiation is in the x-ray wavelength range.
32. The method according to claim 1 wherein said incident electromagnetic radiation is in the gamma ray wavelength range.
33. An apparatus for identifying and sorting selected material objects from a mixture of material objects comprising:
(a) a conveyor that conveys a mixture of material objects into and through an inspection zone;
(b) an electromagnetic radiation source that irradiates said mixture of material objects within said inspection zone with incident electromagnetic radiation;
(c) a sensor that examines said material objects within said inspection zone and measures electromagnetic radiation emanating from said material objects;
(d) a microprocessor computer system that processes said measured measurements of electromagnetic radiation to produce electronic images of said material objects as they pass through said inspection zone and a display coupled to said microprocessor computer system that presents a visual display of said electronic images; and
(e) a computerized pointing device; and
(e)(f) a human operator interface to said visual display such that an operator may selectdesignate selected material objects to be sorted from said mixture of objects using the computerized pointing device; and
(f) an automated device that separates said selected material objects from said mixture of material objects .
34. The apparatus according to claim 33 wherein said mixture of material objects is comprised of solid waste materials.
35. The apparatus according to claim 33 wherein said sensor includes a CCD camera.
36. The apparatus according to claim 33 wherein said conveyor is taken from a group consisting of a belt conveyor, a vibrating pan conveyor, a slide, and a free fall trajectory.
37. The apparatus according to claim 33 wherein said electromagnetic radiation source includes a lamp emitting radiation in the visible range.
38. The apparatus according to claim 33 wherein said display includes a computer monitor.
39. The apparatus according to claim 33 wherein said human operator interface includes a touch sensitive screen.
40. The apparatus according to claim 33 52 wherein said automated device includes an air ejector.
41. The apparatus according to claim 33 52 wherein said automated device includes a robotic arm.
42. The apparatus according to claim 41 wherein said robotic arm includes a suction cup end effector.
43. The apparatus according to claim 42 wherein said suction cup includes means to rotate said suction cup upward and which further includes means for disconnecting vacuum to said suction cup and means to provide compressed air to said suction cup to propel an object held by said suction cup into a receiving chute or bin.
44. The apparatus according to claim 41 wherein said robotic arm includes a driven and retractable spike end effector.
45. The apparatus according to claim 33 wherein said electromagnetic radiation emanating from said material objects as they pass through said inspection zone is passed through a filter to pass only selected wavelengths prior to measuring said electromagnetic radiation.
46. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the microwave wavelength range.
47. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the ultraviolet wavelength range.
48. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the visible wavelength range.
49. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the infrared wavelength range.
50. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the x-ray wavelength range.
51. The apparatus according to claim 33 wherein said electromagnetic radiation source provides incident electromagnetic radiation in the gamma ray wavelength range.
52. The apparatus according to claim 33, further comprising an automated device that separates said selected material objects from said mixture of material objects.
53. The apparatus according to claim 33, wherein the computer system comprises computerized learning and control algorithms so that identifications and selections for the selecting of material items to be sorted is passed from the human operator to the computerized algorithms.
54. The apparatus according to claim 33, wherein the visual display is remote from the inspection zone.
55. The method according to claim 1, wherein step (e) is performed remotely from the location where step (a) is performed.
56. The apparatus according to claim 33, wherein the computer system electronically tracks positions of selected material objects as they are further conveyed.
57. The method according to claim 1, further comprising the step of electronically tracking positions of selected material objects as they are further conveyed.
58. The apparatus according to claim 33, wherein the apparatus visually tags and displays selected material objects.
59. The method according to claim 1, further comprising the step of visually tagging and displaying selected material objects.
US10/252,444 1996-11-04 1997-11-03 Teleoperated robotic sorting system Expired - Lifetime USRE40394E1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/252,444 USRE40394E1 (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US3018396P 1996-11-04 1996-11-04
US10/252,444 USRE40394E1 (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system
PCT/US1997/019680 WO1998019799A1 (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system
US09/297,081 US6124560A (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09297081 Reissue 1997-11-03

Publications (1)

Publication Number Publication Date
USRE40394E1 true USRE40394E1 (en) 2008-06-24

Family

ID=21852943

Family Applications (2)

Application Number Title Priority Date Filing Date
US09/297,081 Ceased US6124560A (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system
US10/252,444 Expired - Lifetime USRE40394E1 (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09/297,081 Ceased US6124560A (en) 1996-11-04 1997-11-03 Teleoperated robotic sorting system

Country Status (3)

Country Link
US (2) US6124560A (en)
AU (1) AU5155798A (en)
WO (1) WO1998019799A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140088765A1 (en) * 2011-04-05 2014-03-27 Zenrobotics Oy Method for invalidating sensor measurements after a picking action in a robot system
CN115072266A (en) * 2022-07-14 2022-09-20 江苏海事职业技术学院 Conveying device for packages and conveying method thereof

Families Citing this family (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1998019799A1 (en) * 1996-11-04 1998-05-14 National Recovery Technologies, Inc. Teleoperated robotic sorting system
BR0013768A (en) * 1999-09-10 2002-04-30 Scanvaegt Int As Method and apparatus for separating supply items for the batch of these items into groups of determined weights
US6859683B2 (en) * 2001-06-04 2005-02-22 Futaba Corporation Of America Integrated robotic cell
US6996456B2 (en) * 2002-10-21 2006-02-07 Fsi International, Inc. Robot with tactile sensor device
US7757863B2 (en) 2003-11-17 2010-07-20 Casella Waste Systems, Inc. Systems and methods for glass recycling at a beneficiator and/or a material recovery facility
US7264124B2 (en) 2003-11-17 2007-09-04 Casella Waste Systems, Inc. Systems and methods for sorting recyclables at a material recovery facility
US7341156B2 (en) * 2003-11-17 2008-03-11 Casella Waste Systems, Inc. Systems and methods for sorting, collecting data pertaining to and certifying recyclables at a material recovery facility
US20060043188A1 (en) * 2004-08-27 2006-03-02 Gregg Kricorissian Imaging method and apparatus for object identification
US7674994B1 (en) * 2004-10-21 2010-03-09 Valerio Thomas A Method and apparatus for sorting metal
CA2565415A1 (en) 2004-11-12 2006-05-26 Casella Waste Systems, Inc. System for and method of mixed-color cullet characterization and certification, and providing contaminant-free, uniformly colored mixed-color cullet
JP4056542B2 (en) * 2005-09-28 2008-03-05 ファナック株式会社 Offline teaching device for robots
WO2007048036A2 (en) * 2005-10-20 2007-04-26 Mtd America, Ltd. Method and apparatus for sorting contaminated glass
US7659486B2 (en) * 2005-10-20 2010-02-09 Valerio Thomas A Method and apparatus for sorting contaminated glass
BRPI0617765A2 (en) * 2005-10-24 2011-08-02 Thomas A Valerio apparatus, system and process for classifying dissimilar materials
US20070174071A1 (en) 2006-01-20 2007-07-26 C&C Acquisition, Llc Techniques for managing information relating to recyclable containers
US20070208455A1 (en) * 2006-03-03 2007-09-06 Machinefabriek Bollegraaf Appingedam B.V. System and a method for sorting items out of waste material
ATE542610T1 (en) * 2006-03-31 2012-02-15 Thomas Valerio METHOD AND DEVICE FOR SORTING FINE NON-FERROUS METALS AND INSULATED WIRE PIECES
AU2008205361B2 (en) * 2007-01-05 2012-06-14 Thomas A. Valerio System and method for sorting dissimilar materials
US8459466B2 (en) * 2007-05-23 2013-06-11 Re Community Energy, Llc Systems and methods for optimizing a single-stream materials recovery facility
FR2923165B1 (en) * 2007-11-07 2014-02-28 Veolia Proprete METHOD AND DEVICE FOR SELECTIVE SORTING OF OBJECTS
US7732726B2 (en) * 2008-04-03 2010-06-08 Valerio Thomas A System and method for sorting dissimilar materials using a dynamic sensor
EP2303462A4 (en) 2008-06-11 2014-01-01 Thomas A Valerio Method and system for recovering metal from processed recycled materials
CA2731506A1 (en) * 2008-07-21 2010-01-28 Mtd America Ltd (Llc) Method and system for removing polychlorinated biphenyls from plastics
EP2424684A4 (en) * 2009-04-28 2014-03-19 Mtd America Ltd Llc Apparatus and method for separating materials using air
GB2471886A (en) * 2009-07-16 2011-01-19 Buhler Sortex Ltd Inspection apparatus
EP2456574A1 (en) * 2009-07-21 2012-05-30 Thomas A. Velerio Method and system for separating and recovering like-type materials from an electronic waste system
US8757523B2 (en) * 2009-07-31 2014-06-24 Thomas Valerio Method and system for separating and recovering wire and other metal from processed recycled materials
AU2010278693A1 (en) * 2009-07-31 2012-03-01 Thomas A. Valerio Method and system for separating and recovering wire and other metal from processed recycled materials
US8809718B1 (en) 2012-12-20 2014-08-19 Mss, Inc. Optical wire sorting
FR3032365B1 (en) 2015-02-10 2017-02-03 Veolia Environnement-VE SELECTIVE SORTING PROCEDURE
FR3032366B1 (en) 2015-02-10 2017-02-03 Veolia Environnement-VE SELECTIVE SORTING PROCESS
US11344036B2 (en) 2015-03-02 2022-05-31 Valka Ehf Apparatus for processing and grading food articles and related methods
US11259531B2 (en) 2015-03-02 2022-03-01 Valka Ehf Apparatus for processing and grading food articles and related methods
US20220061340A1 (en) 2015-03-02 2022-03-03 Valka Ehf Apparatus for processing and grading food articles and related methods
US11278937B2 (en) * 2015-07-16 2022-03-22 Sortera Alloys, Inc. Multiple stage sorting
EP4235540A3 (en) * 2015-09-11 2023-09-27 Berkshire Grey Operating Company, Inc. Robotic systems and methods for identifying and processing a variety of objects
US9937532B2 (en) 2015-12-18 2018-04-10 Berkshire Grey Inc. Perception systems and methods for identifying and processing a variety of objects
US10137573B2 (en) * 2016-08-10 2018-11-27 Cp Manufacturing, Inc. Cleaning robot for recycling equipment
EP3601108A1 (en) 2017-03-20 2020-02-05 Berkshire Grey, Inc. Systems and methods for processing objects including an auto-shuttle system
US11080496B2 (en) 2017-04-18 2021-08-03 Berkshire Grey, Inc. Systems and methods for separating objects using vacuum diverts with one or more object processing systems
US11055504B2 (en) 2017-04-18 2021-07-06 Berkshire Grey, Inc. Systems and methods for separating objects using a vacuum roller with one or more object processing systems
US11205059B2 (en) 2017-04-18 2021-12-21 Berkshire Grey, Inc. Systems and methods for separating objects using conveyor transfer with one or more object processing systems
US11416695B2 (en) 2017-04-18 2022-08-16 Berkshire Grey Operating Company, Inc. Systems and methods for distributing induction of objects to a plurality of object processing systems
US11301654B2 (en) 2017-04-18 2022-04-12 Berkshire Grey Operating Company, Inc. Systems and methods for limiting induction of objects to one or more object processing systems
CN113955367B (en) 2017-04-18 2023-09-12 伯克希尔格雷营业股份有限公司 System and method for processing objects including a space-efficient dispensing station and an automated output process
US11200390B2 (en) 2017-04-18 2021-12-14 Berkshire Grey, Inc. Systems and methods for separating objects using drop conveyors with one or more object processing systems
EP3658600A4 (en) 2017-07-28 2021-06-02 Phillips 66 Company High performance wide-bandgap polymers for organic photovoltaics
EP3684521B1 (en) * 2017-09-19 2024-04-24 Valka Ehf Apparatus for processing and grading food articles and related method
SE544077C2 (en) * 2018-04-22 2021-12-14 Zenrobotics Oy Waste Sorting Robot Gripper
SE543130C2 (en) 2018-04-22 2020-10-13 Zenrobotics Oy A waste sorting robot gripper
SE544741C2 (en) 2018-05-11 2022-11-01 Genie Ind Bv Waste Sorting Gantry Robot and associated method
JP7124509B2 (en) * 2018-07-19 2022-08-24 オムロン株式会社 SIMULATION DEVICE, SIMULATION PROGRAM AND SIMULATION METHOD
CN109261539B (en) * 2018-08-17 2021-06-18 湖北文理学院 Garbage sorting system and method based on visual identification and convolutional neural network
CN112770851B (en) * 2018-09-18 2022-09-13 安普机器人技术公司 Vacuum extraction for material sorting applications
CA3126766C (en) 2018-10-23 2023-09-19 Berkshire Grey, Inc. Systems and methods for dynamic processing of objects with data verification
CN112930547A (en) 2018-10-25 2021-06-08 伯克希尔格雷股份有限公司 System and method for learning extrapolated optimal object transport and handling parameters
NL2022319B1 (en) 2018-12-28 2020-07-23 Bollegraaf Patents And Brands B V Suction unit for a robotic sorter
CA3128185A1 (en) 2019-01-31 2020-08-06 Four Growers, Inc. Crop harvesting robot
GB201906418D0 (en) 2019-05-07 2019-06-19 Valka Ehf Conveyor system and method
CN110137117B (en) * 2019-06-05 2024-01-30 上海赢朔电子科技股份有限公司 Semiconductor double-rail DD motor high-speed automatic test coding and taping separator
US20220241823A1 (en) 2019-07-29 2022-08-04 Industries Machinex System for sorting materials with vacuum apparatus
WO2021030339A1 (en) * 2019-08-15 2021-02-18 Qelzal Corporation Dba Kelzal Method and system for efficient multi-modal object recognition
CN110861853B (en) * 2019-11-29 2021-10-19 三峡大学 Intelligent garbage classification method combining vision and touch
US11753257B2 (en) 2019-12-16 2023-09-12 AMP Robotics Corporation Bidirectional air conveyor device for material sorting and other applications
WO2021126879A1 (en) 2019-12-16 2021-06-24 AMP Robotics Corporation A suction gripper cluster device for material sorting and other applications
EP4076870A4 (en) * 2019-12-16 2024-04-17 Amp Robotics Corp An actuated air conveyor device for material sorting and other applications
CN111626117B (en) * 2020-04-22 2023-04-18 杭州电子科技大学 Garbage sorting system and method based on target detection
GB2612653A (en) * 2021-11-09 2023-05-10 Danu Robotics Ltd Robotic system and method for picking and/or sorting objects
EP4201600A1 (en) * 2021-12-21 2023-06-28 Open Mind Ventures, S.L.U. System for sorting objects comprising a plurality of catching elements

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4805778A (en) * 1984-09-21 1989-02-21 Nambu Electric Co., Ltd. Method and apparatus for the manipulation of products
US4942538A (en) * 1988-01-05 1990-07-17 Spar Aerospace Limited Telerobotic tracker
US5044002A (en) * 1986-07-14 1991-08-27 Hologic, Inc. Baggage inspection and the like
US5141110A (en) * 1990-02-09 1992-08-25 Hoover Universal, Inc. Method for sorting plastic articles
US5260576A (en) * 1990-10-29 1993-11-09 National Recovery Technologies, Inc. Method and apparatus for the separation of materials using penetrating electromagnetic radiation
US5299693A (en) * 1991-04-12 1994-04-05 Ubaldi Richard A Method and apparatus for extracting selected materials
US5411147A (en) * 1993-01-28 1995-05-02 Bond; David S. Dynamic landfill recycling system
US5423431A (en) * 1989-03-23 1995-06-13 Sellsberg Engineering Ab Method and an apparatus for waste handling
US5520290A (en) * 1993-12-30 1996-05-28 Huron Valley Steel Corporation Scrap sorting system
US5555984A (en) * 1993-07-23 1996-09-17 National Recovery Technologies, Inc. Automated glass and plastic refuse sorter
US5600700A (en) * 1995-09-25 1997-02-04 Vivid Technologies, Inc. Detecting explosives or other contraband by employing transmitted and scattered X-rays
US5699400A (en) * 1996-05-08 1997-12-16 Vivid Technologies, Inc. Operator console for article inspection systems
US5752607A (en) * 1996-03-18 1998-05-19 Moen Incorporated Process for distinguishing plumbing parts by the coatings applied thereto
US6124560A (en) * 1996-11-04 2000-09-26 National Recovery Technologies, Inc. Teleoperated robotic sorting system

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4805778A (en) * 1984-09-21 1989-02-21 Nambu Electric Co., Ltd. Method and apparatus for the manipulation of products
US5044002A (en) * 1986-07-14 1991-08-27 Hologic, Inc. Baggage inspection and the like
US4942538A (en) * 1988-01-05 1990-07-17 Spar Aerospace Limited Telerobotic tracker
US5423431A (en) * 1989-03-23 1995-06-13 Sellsberg Engineering Ab Method and an apparatus for waste handling
US5141110A (en) * 1990-02-09 1992-08-25 Hoover Universal, Inc. Method for sorting plastic articles
US5339962A (en) * 1990-10-29 1994-08-23 National Recovery Technologies, Inc. Method and apparatus for sorting materials using electromagnetic sensing
US5260576A (en) * 1990-10-29 1993-11-09 National Recovery Technologies, Inc. Method and apparatus for the separation of materials using penetrating electromagnetic radiation
US5299693A (en) * 1991-04-12 1994-04-05 Ubaldi Richard A Method and apparatus for extracting selected materials
US5411147A (en) * 1993-01-28 1995-05-02 Bond; David S. Dynamic landfill recycling system
US5555984A (en) * 1993-07-23 1996-09-17 National Recovery Technologies, Inc. Automated glass and plastic refuse sorter
US5520290A (en) * 1993-12-30 1996-05-28 Huron Valley Steel Corporation Scrap sorting system
US5600700A (en) * 1995-09-25 1997-02-04 Vivid Technologies, Inc. Detecting explosives or other contraband by employing transmitted and scattered X-rays
US5752607A (en) * 1996-03-18 1998-05-19 Moen Incorporated Process for distinguishing plumbing parts by the coatings applied thereto
US5699400A (en) * 1996-05-08 1997-12-16 Vivid Technologies, Inc. Operator console for article inspection systems
US6124560A (en) * 1996-11-04 2000-09-26 National Recovery Technologies, Inc. Teleoperated robotic sorting system

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140088765A1 (en) * 2011-04-05 2014-03-27 Zenrobotics Oy Method for invalidating sensor measurements after a picking action in a robot system
CN115072266A (en) * 2022-07-14 2022-09-20 江苏海事职业技术学院 Conveying device for packages and conveying method thereof

Also Published As

Publication number Publication date
AU5155798A (en) 1998-05-29
US6124560A (en) 2000-09-26
WO1998019799A1 (en) 1998-05-14

Similar Documents

Publication Publication Date Title
USRE40394E1 (en) Teleoperated robotic sorting system
US5299693A (en) Method and apparatus for extracting selected materials
US11494575B2 (en) Systems and methods for identifying and processing a variety of objects
US11351575B2 (en) Perception systems and methods for identifying and processing a variety of objects
US20190130560A1 (en) Systems and methods for optical material characterization of waste materials using machine learning
US20200342240A1 (en) Systems and methods for detecting waste receptacles using convolutional neural networks
JP3484196B2 (en) Method and apparatus for sorting material parts
US20060273257A1 (en) Computer assisted bag screening system
AU2016200745A1 (en) Selective sorting method
US20230365352A1 (en) Bidirectional air conveyor device for material sorting and other applications
EP3866988B1 (en) System for the sorting of materials from collected waste
CN113183138A (en) Garbage carrying and sorting robot and control method thereof
CN112090782A (en) Man-machine cooperative garbage sorting system and method
RU2624288C1 (en) Method of sorting of waste
DE19614336C1 (en) Operating field for input and control of industrial robots or manipulators
CN110914021A (en) Operating device with an operating device for carrying out at least one work step, and method and computer program
RU2806224C1 (en) Waste sorting method
Lemeshko et al. A rational way of sorting municipal solid waste
DE10217536A1 (en) Sorting valuable components from different or pre-sorted material, comprises irradiating the material with light and analyzing the reflected light optoelectronically

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FPAY Fee payment

Year of fee payment: 12

AS Assignment

Owner name: CALTIUS PARTNERS III, LP, AS AGENT, CALIFORNIA

Free format text: SECURITY AGREEMENT;ASSIGNOR:NATIONAL RECOVERY TECHNOLOGIES, INC.;REEL/FRAME:028130/0530

Effective date: 20120426

AS Assignment

Owner name: NATIONAL RECOVERY TECHNOLOGIES, LLC, TENNESSEE

Free format text: CHANGE OF NAME;ASSIGNOR:NATIONAL RECOVERY TECHNOLOGIES, INC.;REEL/FRAME:034995/0458

Effective date: 20120628

AS Assignment

Owner name: NATIONAL RECOVERY TECHNOLOGIES, LLC, TENNESSEE

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CALTIUS PARTNERS III, LP;REEL/FRAME:045293/0583

Effective date: 20180320

AS Assignment

Owner name: TRUE WEST CAPITAL PARTNERS FUND II, LP, OREGON

Free format text: SECURITY INTEREST;ASSIGNOR:NATIONAL RECOVERY TECHNOLOGIES, LLC;REEL/FRAME:045353/0387

Effective date: 20180326

AS Assignment

Owner name: PNC BANK, NATIONAL ASSOCIATION, PENNSYLVANIA

Free format text: SECURITY INTEREST;ASSIGNORS:EMERGING ACQUISITIONS, LLC;NATIONAL RECOVERY TECHNOLOGIES, LLC;NIHOT RECYCLING TECHNOLOGY B.V.;AND OTHERS;REEL/FRAME:049513/0198

Effective date: 20190501