US20130205920A1 - Automated visual pipetting - Google Patents

Automated visual pipetting Download PDF

Info

Publication number
US20130205920A1
US20130205920A1 US13/763,715 US201313763715A US2013205920A1 US 20130205920 A1 US20130205920 A1 US 20130205920A1 US 201313763715 A US201313763715 A US 201313763715A US 2013205920 A1 US2013205920 A1 US 2013205920A1
Authority
US
United States
Prior art keywords
pipetting
work surface
control unit
tool head
command unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/763,715
Inventor
Adam Perry Tow
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US13/763,715 priority Critical patent/US20130205920A1/en
Publication of US20130205920A1 publication Critical patent/US20130205920A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B01PHYSICAL OR CHEMICAL PROCESSES OR APPARATUS IN GENERAL
    • B01LCHEMICAL OR PHYSICAL LABORATORY APPARATUS FOR GENERAL USE
    • B01L3/00Containers or dishes for laboratory use, e.g. laboratory glassware; Droppers
    • B01L3/02Burettes; Pipettes
    • B01L3/021Pipettes, i.e. with only one conduit for withdrawing and redistributing liquids
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B01PHYSICAL OR CHEMICAL PROCESSES OR APPARATUS IN GENERAL
    • B01LCHEMICAL OR PHYSICAL LABORATORY APPARATUS FOR GENERAL USE
    • B01L3/00Containers or dishes for laboratory use, e.g. laboratory glassware; Droppers
    • B01L3/02Burettes; Pipettes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/10Processes of additive manufacturing
    • B29C64/171Processes of additive manufacturing specially adapted for manufacturing multiple 3D objects
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/20Apparatus for additive manufacturing; Details thereof or accessories therefor
    • B29C64/205Means for applying layers
    • B29C64/209Heads; Nozzles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/20Apparatus for additive manufacturing; Details thereof or accessories therefor
    • B29C64/295Heating elements
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/30Auxiliary operations or equipment
    • B29C64/386Data acquisition or data processing for additive manufacturing
    • B29C64/393Data acquisition or data processing for additive manufacturing for controlling or regulating additive manufacturing processes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B33ADDITIVE MANUFACTURING TECHNOLOGY
    • B33YADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
    • B33Y30/00Apparatus for additive manufacturing; Details thereof or accessories therefor
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N35/00Automatic analysis not limited to methods or materials provided for in any single one of groups G01N1/00 - G01N33/00; Handling materials therefor
    • G01N35/10Devices for transferring samples or any liquids to, in, or from, the analysis apparatus, e.g. suction devices, injection devices
    • G01N35/1009Characterised by arrangements for controlling the aspiration or dispense of liquids
    • G01N35/1011Control of the position or alignment of the transfer device
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/10Processes of additive manufacturing
    • B29C64/106Processes of additive manufacturing using only liquids or viscous materials, e.g. depositing a continuous bead of viscous material
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B29WORKING OF PLASTICS; WORKING OF SUBSTANCES IN A PLASTIC STATE IN GENERAL
    • B29CSHAPING OR JOINING OF PLASTICS; SHAPING OF MATERIAL IN A PLASTIC STATE, NOT OTHERWISE PROVIDED FOR; AFTER-TREATMENT OF THE SHAPED PRODUCTS, e.g. REPAIRING
    • B29C64/00Additive manufacturing, i.e. manufacturing of three-dimensional [3D] objects by additive deposition, additive agglomeration or additive layering, e.g. by 3D printing, stereolithography or selective laser sintering
    • B29C64/10Processes of additive manufacturing
    • B29C64/106Processes of additive manufacturing using only liquids or viscous materials, e.g. depositing a continuous bead of viscous material
    • B29C64/118Processes of additive manufacturing using only liquids or viscous materials, e.g. depositing a continuous bead of viscous material using filamentary material being melted, e.g. fused deposition modelling [FDM]
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B33ADDITIVE MANUFACTURING TECHNOLOGY
    • B33YADDITIVE MANUFACTURING, i.e. MANUFACTURING OF THREE-DIMENSIONAL [3-D] OBJECTS BY ADDITIVE DEPOSITION, ADDITIVE AGGLOMERATION OR ADDITIVE LAYERING, e.g. BY 3-D PRINTING, STEREOLITHOGRAPHY OR SELECTIVE LASER SINTERING
    • B33Y50/00Data acquisition or data processing for additive manufacturing
    • B33Y50/02Data acquisition or data processing for additive manufacturing for controlling or regulating additive manufacturing processes

Definitions

  • the present invention relates to automated visual pipetting (AVP) using a camera and automation procedures using techniques available in three dimensional fabrication systems.
  • AVP automated visual pipetting
  • a pipette (also called a pipet, pipettor, or chemical dropper) is a laboratory tool used to transport a measured volume of liquid.
  • Pipettes are commonly used in molecular biology, analytical chemistry, and medical tests. Pipettes come in several designs for various purposes with differing levels of accuracy and precision, from single piece glass pipettes to more complex adjustable or electronic pipettes. Many pipette types work by creating a partial vacuum above the liquid-holding chamber and selectively releasing this vacuum to draw up and dispense liquid.
  • the shortcomings of the prior art can be overcome and additional advantages can be provided with the Automated Visual Pipetting (“AVP”) systems and techniques described herein.
  • the present invention can improve the accuracy, ease of use and efficiency of pipetting procedures in order to achieve a drastic improvement in performance and quality.
  • Embodiments of the present invention involving Automated Visual Pipetting are conceived with a few goals in mind, most notably: (1) to be visually consistent with the logic of experiments performed by hand; (2) such that a first day undergraduate research associate could be capable of operating such an embodiment, and running experiments within minutes of engaging it; and (3) able to be added to machines that may have non-pipetting uses, such as three dimensional fabricators (i.e., additive manufacturing) or other computer numerical control (CNC) machine tools like milling.
  • three dimensional fabricators i.e., additive manufacturing
  • CNC computer numerical control
  • a camera is mounted on a pipette or another type of tool that may be a deposition or non-deposition tool (e.g., milling).
  • a camera feed and recognition software can enable users to replace existing micro-pipetting techniques with a computerized process that may be controlled with a few simple mouse clicks, while the user can directly visualize a live experimental setup.
  • the embodiment also allows for integration with common molecular biology procedure “kits,” and enables the process to be automated and visualized without requiring constant interaction by the user.
  • Such embodiments of the present invention may be uniquely capable of pipetting and automating cell culture using a single machine and essentially the same software.
  • embodiments of the present invention as disclosed herein can allow for cells to be deposited in exact (or very near exact) geometric patterns, providing several new opportunities such as optimizing repeatable cell line-specific deposition patterns for achieving desired confluence in a specific time frame, and experimenting using novel geometric cell arrangements which could potentially include platforms for studying single neuron synapses or creating entire “manufactured” organ systems.
  • Embodiments of the present invention may involve a camera and pipetting tool that, for example, is guided by a control unit receiving instructions from a fabrication and/or pipetting command unit (which may be a computer) running either a locally-stored or server-based fabrication and/or pipetting software application.
  • the computer may be integrated into a fabrication and/or pipetting device, or connected to the fabrication and/or pipetting device via a wireless connection such as Bluetooth, WLAN, NFC or other wireless communication technologies, or a wired connection such as Ethernet, USB, FireWire, serial or parallel connection, or other wired communication technologies.
  • a three dimensional pipetting device having a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head, and a plurality of interchangeable pipetting tips that can be affixed to the pipetting tool head, such that the control unit can operate the pipetting tool head to selectively use one of the plurality of interchangeable pipetting tips, and selectively draw up or dispense liquid from or to one or more containers on a work surface.
  • a three dimensional pipetting device having a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head, and a camera for recording data including the position of one or more containers on a work surface, such that the control unit can operate the pipetting tool head to selectively draw up or dispense liquid from or to the one or more containers on the work surface.
  • the three dimensional pipetting device may also be a three dimensional fabricating system.
  • the pipetting command unit is configured to receive and process data recorded by the camera to determine the location and size of at least one container on the work surface.
  • the pipetting command unit is configured to receive and process data recorded by the camera to generate a digital image of a plurality of items on the work surface, which may optionally be output to an external monitor to display a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface. Additionally, the pipetting command unit is configured to receive data recorded by the camera and visually simulate the performance of a procedure that can be performed by the pipetting tool head. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera to identity an item on the work surface. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively draw up liquid from at least one container on the work surface. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively deposit liquid into at least one container on the work surface.
  • a method for using a three dimensional pipetting device including the steps of transmitting instructions for operating a camera from a pipetting command unit to a control unit, operating the camera with the control unit to record image data of one or more containers on a work surface, transmitting recorded image data from the control unit to the pipetting command unit, transmitting instructions for operating a pipetting tool head from a pipetting command unit to a control unit; and operating the pipetting tool head to selectively draw up liquid from the one or more containers on the work surface.
  • the pipetting command unit processes recorded image data and determines the location of at least one container on the work surface.
  • the pipetting command unit processes recorded image data and generates a digital image of a plurality of items on the work surface, and the digital image may include a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface. Additionally, the pipetting command unit processes recorded image data received from the control unit and simulates performance of a procedure that can be performed by the pipetting tool head. Additionally, the pipetting command unit processes recorded image data and detects a visual indicator (that may be a QR code) in at least one of the recorded images that identifies an item on the work surface.
  • a visual indicator that may be a QR code
  • the pipetting command unit processes recorded image data, detects a visual indicator (that may be a QR code) in at least one of the recorded images as an instruction to use the pipetting tool head, and selectively draws up or deposits liquid from or into at least one container on the work surface in response to said instruction. Additionally, the pipetting command unit processes recorded image data received from the control unit and detects whether a prescribed amount of liquid has been aspirated from the one or more containers on the work surface by the pipetting tool head.
  • a visual indicator that may be a QR code
  • One great benefit of the present invention is that many embodiments of Automated Visual Pipetting can be combined with, and implemented in, three dimensional fabricators (3D printers).
  • three dimensional fabricators 3D printers
  • many embodiments of the present invention described herein relate to a joint three dimensional fabricator and Automated Visual Pipetting device, it should be readily apparent to the reader that stand-alone Automated Visual Pipetting devices are within the scope of the present invention, and that a three dimensional fabricator is not required to implement the present invention.
  • the present invention has many embodiments, some of which are described herein, and others which should be apparent to the reader or inferred from what is taught herein.
  • FIG. 1 is a perspective view of a three dimensional fabricator.
  • FIG. 2 shows an embodiment of the invention in which a camera is affixed to the deposition tool head.
  • FIG. 3 shows an embodiment of the invention in which a visual code is used by the on-board camera to guide motion and interchange tips, or identify items on the machine's build-tray.
  • FIG. 4 is a perspective view of a three dimensional pipetting device in accordance with an embodiment of the invention.
  • FIG. 1 provides a perspective view of a prior art three dimensional fabricating system.
  • Fabrication system 100 includes fabricator 101 with material deposition tool head 102 (also referred to herein as deposition tool or deposition head), control unit 103 having one or more actuators and sensors configured to control operating characteristics of material deposition tool 102 , and build tray (i.e., build surface) 104 .
  • Fabrication command unit 105 may be coupled to fabricator 101 as a component physically inside fabricator 101 , or it may be coupled as an external device (e.g., computer) via a wired or wireless connection.
  • Fabrication command unit 105 includes processor 106 , memory 107 , and fabrication software application 108 that can be stored in memory 107 and executed by processor 106 . It should be appreciated that control unit 103 of fabricator 101 may be configured to receive instructions from fabrication command unit 105 such that fabricator 101 can fabricate an output product on build surface 104 from materials dispensed by material deposition tool 102 . Fabrication software application 108 can generate tool path information for fabricator 101 and delineate how material can be used to generate shapes. Complex CAD programs may also be used to generate the intended geometry.
  • Embodiments of the present invention may be implemented in any three dimensional fabricating system (i.e., additive manufacturing device or 3D Printer), for example, as illustrated in FIG. 1 and described above, that is suitable for performing Automated Visual Pipetting techniques.
  • exemplary three dimensional fabricating system or components thereof that may be suitable for Automated Visual Pipetting are described in U.S. Pub. No. 2012/0241993 entitled “SYSTEMS AND METHODS FOR FREEFORM FABRICATION OF FOAMED STRUCTURES” and published on Sept. 27, 2012 (filed as U.S. application Ser. No. 13/356,194 on Jan. 23, 2012) and U.S. Pat. No. 7,625,198 to Lipson et al.
  • Material deposition tool 102 may include a mounted pipetting tool and camera, or be replaced with a pipetting tool and camera.
  • material deposition tool 102 may include a mounted pipetting tool or be replaced with a pipetting tool, and one or more cameras may be mounted on other surfaces of fabricator 101 .
  • Control unit 103 having one or more actuators and sensors configured to control operating characteristics of material deposition tool 102 , may similarly be configured to control operating characteristics of a pipetting tool and one or more cameras.
  • Experiments to undergo pipetting procedures may be placed onto build tray (i.e., build surface) 104 .
  • Fabrication command unit 105 may be configured to support manual and automated use of a pipetting tool and one or more cameras.
  • fabrication software application 108 may include or be replaced with Automated Visual Pipetting software to receive pipetting and/or camera requests from a user, generate pipetting tool and/or camera path information, and direct fabricator 101 and perform manual and automated pipetting procedures, as well as operate the one or more cameras.
  • Automated Visual Pipetting software included with, or replacing, fabrication software application 108 can be stored in memory 107 and executed by processor 106 .
  • control unit 103 of fabricator 101 may be configured to receive instructions from fabrication command unit 105 such that fabricator 101 can conduct pipetting procedures on experiments placed upon build surface 104 from a pipetting tool mounted on, or replacing, deposition tool 102 .
  • fabrication system 100 may be referred to as AVP system 100
  • fabrication command unit 105 may be referred to as AVP command unit 105
  • fabricator 101 may be referred to as a three dimensional pipetting device, as it supports three dimensional movement of the pipetting tool above build surface 104 .
  • Embodiments of the present invention may be implemented in three dimensional fabricating systems (i.e., additive manufacturing devices or 3D Printers), similar to the example illustrated in FIG. 1 and described above.
  • U.S. Pub. No. 2012/0241993 entitled “SYSTEMS AND METHODS FOR FREEFORM FABRICATION OF FOAMED STRUCTURES” and published on Sep. 27, 2012 (filed as U.S. application Ser. No. 13/356,194 on Jan. 23, 2012) provides a similar depiction of FIG. 1 with an accompanying disclosure.
  • additional information regarding many components of three dimensional fabricating systems may be found in U.S. Pat. No. 7,625,198 to Lipson et al.
  • tool head 201 can accept a plurality of disposable (or reusable) pipette tips (or needles) 202 and act as a pipette or micropipette would function in the hand of a biologist.
  • Camera unit 203 can send a live video or image feed of lens 204 to a computer (e.g., Fabrication command unit 105 of FIG. 1 ).
  • the live video or image feed can be used by the computer to map out an experimental setup placed on a build surface (e.g., build tray 104 of FIG. 1 ) which is accessible to tool head 201 .
  • Tool head 201 may be mounted to, or replace, material deposition tool 102 of FIG. 1 .
  • Automated Visual Pipetting software can assess where items are in a setup by visualizing items, the particular containers being used, or by other visual indicators, such as a particular shape, a QR code, a bar code, one or more numbers and/or letters, a specific packaging design or logo, etc.
  • Camera 203 may be used by AVP system 100 to automatically guide tool head 201 , or to assess distance, for example by measuring pixels between two points of known distance and thereby calculating a distance from those points by how big they appear to the camera (on a pixel basis). Camera 203 may be able to visualize multiple resolutions, by the use of digital or optical zoom.
  • Camera 203 could be calibrated to detect different colored circumferences on various types of containers, to easily identify components of a commonly used kit of experimental materials.
  • Automated Visual Pipetting software e.g., included with, or replacing, fabrication software application 108
  • Automated Visual Pipetting software can assess what pipetting operations to conduct with respect to particular items detected by the camera (e.g., drawing liquids or depositing liquids) by visualizing instructions on, e.g., a container or build tray, such as by recognizing the particular containers being used, or by other visual indicators, such as a particular shape, a QR code, a bar code, one or more numbers and/or letters, a specific packaging design or logo, etc.
  • the visual indicator may directly identify the procedure to be performed; which, with the use of a kit, may require the user to identify a substrate on which to perform the procedure.
  • Automated Visual Pipetting software may read instructions directly from the items, or suggest (and execute) possible procedures based on the identity of items on the tray and their corresponding potential uses in a series of known procedures.
  • Camera 203 may allow a user viewing a live video or image feed from Camera 203 via the Automated Visual Pipetting software to guide tool head 201 with, e.g., computer commands, a mouse or a joystick.
  • the Automated Visual Pipetting software may be designed to limit the actions of a user in order to prevent errors or unintended consequences resulting from user mistakes.
  • the software may prevent the user from initiating a pipetting action when tool head 201 is not aligned properly over the target to be pipetted.
  • the software may prevent the user from initiating a pipetting action when tool head 201 is aligned over a sample that should not be pipetted (or has already been pipetted).
  • Embodiments of the present invention may cover the entire spectrum of AVP functionality, with embodiments on one end of the spectrum completely automating the described procedures, and embodiments at the other end allowing these procedures to be manually directed by a user of the computer running the Automated Visual Pipetting software.
  • a single device may be configurable such that it can implement a number of different embodiments, each one involving different levels of automation and user interaction.
  • FIG. 3 illustrates additional aspects of an embodiment of the present invention.
  • Atop build table 301 e.g., build tray 104 of FIG. 1
  • a plurality of kits and accessories can be placed, such as the standard pipette tip box 305 .
  • Tip box 305 has several holes 306 in which pipette tips 309 are removed by tool head 201 of FIG. 2 (as illustrated at the top of FIG. 3 ).
  • a camera e.g., Camera 203 of FIG. 2
  • AVP system 100 can determine the distance of tool head 201 from box 305 by the relative (pixel) size of target 308 .
  • AVP system 100 can measure the distance in pixels between target rings 303 and 304 to determine the distance between tool head 201 and box 305 or build table 301 . Having target rings 303 and 304 on build table 301 , is one possible way to eliminate the need for target 308 on box 305 , since it has a feature or object of known dimension.
  • QR code 307 or QR code 302 can be used to identify a particular kit (e.g., box 305 ) or build table 301 , respectively.
  • the distance can be calculated using target 308 , it can also be calculated simply by using the relative (pixel) size of a pipette tip's visible, proximal lumen and comparing it to the known size of a pipette tip's visible lumen in a tip box.
  • the same technique can be used to determine the distance from any other object, such as a conical tube, which could be visually matched by the Automated Visual Pipetting software to a likely, known tube and then distance measured by the relative pixel measuring system just described. Even if the tube were not directly in line below a camera, Automated Visual Pipetting software can be adjusted to compensate by noting the difference in appearance of a tube that was perfectly flat.
  • a camera could also be used to detect the positioning of material inside a vessel (e.g. conical tube), and then position the pipette tip to capture that material, and also measure the volume of material aspirated (e.g. by visualizing the extent to which the pipette tip has been filled) to assure it was successful. (Such visualization may require the use or a secondary camera or mirror system to the primary camera to visualize the engaged deposition pipette tip.)
  • a jet of aspirated air will cause the liquid to splash inside the pipette tip.
  • the knowledge of the fluid properties e.g.
  • the present invention allows the present system to predict the level at which the properly aspirated amount of fluid should rise to.
  • the present invention can discern whether a uniform aspiration was carried out, (e.g. was there a jet of air caused by lack of material) or whether liquid rises to the proper level in the tip (e.g. as in cases both where there was or was not an air jet, as might be the case for a viscous fluid).
  • the software may use a comparative algorithm between an empty tip and the aspirated state (or expectations thereof) to perform the above functionality.
  • Laboratory procedures using an embodiment of the present invention may be performed in a variety of ways: by using the live video or image feed, tool head 201 could be controlled by simple mouse clicks (or other input methods) on a connected computing device, which may be connected directly, or over a local wireless connection, or via the Internet. These techniques could replace the need for skilled pipetting procedures by providing similar results with as little as a few computer commands (e.g., a few mouse clicks), allowing a live experiment to be run by a user without special training or manual skills. Moreover, these techniques could be initiated with a virtual run through, and then performed without requiring any user interaction.
  • a user could make a few selections on a computer to initiate an experiment, the computer could visually present a virtual demonstration of that procedure, await for user confirmation to begin, and then perform the process without requiring further user interaction.
  • a procedure can be programmed for a particular experimental setup.
  • the experimenter would simply have to place the appropriate components, kits, and accessories on a build table, and AVP system 100 would recognize these items and simply proceed with performing an experiment based on a preprogrammed set of instructions. These instructions need not be relative to the position of items on the build table, but rather to what the items actually are, which is now possible due to the use of a video camera.
  • AVP system 100 could identify the various components, or allow the user to do so, and then proceed with the experiment as would a person, being able to adapt to the “random” positioning of objects by knowing what and where things are, not just where they should be.
  • inventions can also automate the use of cell cultures by processing cell growth data and depositing cells in geometries which will lead to a particular result, such as a desired cell confluence (in x days), or yield a particular growth pattern, like a row of neurons.
  • the embodiments can run as follows. Several items can be placed on a build tray, they will each be digitally identified either by reference to a database of known items, by a scanned code, or by object properties (e.g., opening size, position), all of which can use digital mapping via the camera. Each item can then be superimposed with a digital outline shape, showing the computer's recognition of an opening, and either identifying what an object is or allowing the object's identity to be assigned. With objects known, a pre-programmed protocol can be performed. Alternatively, placing the items on a tray can be a pre-defined indicator for what protocol the computer should run.
  • object properties e.g., opening size, position
  • a series of selections on a computer can be made to indicate the pipetting action to be taken. This can be performed either in real time or be input into a computer, optionally previewed in a virtual run through, and then performed AVP system 100 will take appropriate measurements to assure it pipettes correct volume, as well as use the camera to help identify optimal pipette placement in a container and the amount of fluid in it (recall the z-axis is virtually a constant).
  • Embodiments can also involve creating a custom build tray (e.g., using a plastic deposition tool) and then setting up a unique experimental setup in that tray. After use, the tray can be discarded, replaced, cleaned or reused.
  • a custom build tray e.g., using a plastic deposition tool
  • Embodiments could take a crowded build (work) tray with many tube trays and other components on it, and then visually separate those components out when displaying them on a computer screen (for example by identifying component edges or small areas of unused space), so that they are easier to visualize for the user, and not crowded in a displayed video feed despite their physical positioning.
  • This can allow more efficient utility of build tray space, and potentially allow for more logical click-through protocol setup.
  • a user could, for example, digitally rearrange the position of components as they appear on the computer screen (versus how they actually sit on the tray), and customize the visualization describing data about those components on the same screen.
  • Embodiments could also track the contents of each individual component over a series of experiments. So, for example, a QR code (or otherwise visually) labeled 12 -well plate could have reagent X added to it on day one, incubated overnight, and then upon replacement onto the build (work) tray on day two, the camera would check the plate data against a database containing recorded actions of AVP system 100 on the previous day, such that the user would be aware that reagent X was added to particular wells on the plate the previous day.
  • a series of tubes could have labels in several areas, for example the cap. In such a case, the user may wish to leave the closed tubes on the tray, allow the camera to identify the codes, and remove and replace one tube at a time (to reduce error) and then open them.
  • Embodiments can also contain additional cameras at various positions (and/or with various angles) on one or more tool heads or on the device frame itself for a variety of advances uses, including reading identification codes not easily identified from above the build (work) tray.
  • Embodiments of the present invention could also be used with an integrated label gun, which would add labels to components in an experiment as needed.
  • automation and manufacturing tasks can be more precisely controlled and audited.
  • a camera can be used to verify the quality of three dimensional items as they are being printed, or precisely pick and place, or pipette, deposition tasks.
  • FIG. 4 illustrates one possible embodiment with a modified version of Fabrication system 100 as shown in FIG. 1 and described herein.
  • AVP system 400 can perform Automated Visual Pipetting.
  • Pipetting tool and camera 409 (such as the pipetting tool and camera illustrated in FIG. 2 and described herein) is mounted for use on deposition tool head 402 .
  • Control unit 403 has one or more actuators and sensors configured to control operating characteristics of pipetting tool and camera 40 ⁇ .
  • Experiments to undergo pipetting procedures may be placed onto work tray (i.e., work surface) 404 .
  • Container 410 may be a beaker or petri dish and contain an item such as a liquid, substrate or cell culture.
  • Pipette tip box 411 (such as the pipette tip box illustrated in FIG. 3 and described herein) may also be placed on work tray 404 within reach of pipetting tool and camera 409 .
  • AVP command unit 405 may be configured to support manual and automated use of pipetting tool and camera 409 .
  • AVP software application 408 may receive pipetting and/or camera requests from a user, generate pipetting tool and/or camera path information, and direct control unit 403 to perform manual and automated pipetting procedures, as well as operate one or more cameras.
  • Automated Visual Pipetting software included can be stored in memory 407 and executed by processor 406 .
  • control unit 403 of AVP system 400 may be configured to receive instructions from AVP command unit 405 such that AVP system 400 can conduct pipetting procedures on experiments placed upon work surface 404 from pipetting tool and camera 409 mounted on deposition tool 402 . Therefore, in various embodiments of the present invention, AVP system 400 may be referred to as a three dimensional pipetting device because it supports two axis movement of the pipetting tool above work surface 404 , as well as up and down movement.
  • AVP command unit 405 may have a wireless or wired connection to external computer screen (i.e., monitor) 413 direct output of recorded images from pipetting tool and camera 409 to external computer screen 413 , such as to display image 415 of container 410 , and image 416 of pipette tip box 411 .
  • external computer screen i.e., monitor
  • a computer screen can also be used in a variety of way to take advantage of the benefits offered by embodiments of the present invention.

Landscapes

  • Chemical & Material Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Materials Engineering (AREA)
  • Physics & Mathematics (AREA)
  • Manufacturing & Machinery (AREA)
  • Mechanical Engineering (AREA)
  • Optics & Photonics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Clinical Laboratory Science (AREA)
  • Chemical Kinetics & Catalysis (AREA)
  • Microscoopes, Condenser (AREA)

Abstract

A system and method is described for Automated Visual Pipetting on a machine such as a three dimensional fabrication device like a 3D Printer or other computer numerical control (CNC) machine tools, for improved speed, accuracy and reliability in pipetting procedures. A camera is mounted on a pipette or another type of tool that may be a deposition or non-deposition tool (e.g., milling). A camera feed and recognition software can enable users to replace existing micro-pipetting techniques with a computerized process that may be controlled with a few simple mouse clicks, while the user can directly visualize a live experimental setup. The embodiment also allows for integration with common molecular biology procedure “kits,” and enables the process to be automated and visualized without requiring constant interaction by the user.

Description

    REFERENCE TO RELATED APPLICATIONS
  • This applicant claims the benefit of U.S. Provisional Application Ser. No. 61/633,433, filed Feb. 10, 2012 and incorporated by reference herein; U.S. Provisional Application Ser. No. 61/741,368, filed Jul. 18, 2012 and incorporated by reference herein; U.S. Provisional Application Ser. No. 61/689,963, filed Jun. 18, 2012 and incorporated by reference herein.
  • Co-pending patent application no. 13/761,272 entitled “MULTI-AXIS, MULTI-PURPOSE ROBOTICS AUTOMATION AND QUALITY ADAPTIVE ADDITIVE MANUFACTURING” filed on Feb. 7, 2013 having named inventor Adam Perry Tow is hereby incorporated by reference in its entirety and for all purposes.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to automated visual pipetting (AVP) using a camera and automation procedures using techniques available in three dimensional fabrication systems.
  • 2. Background
  • There have been many developments in additive manufacturing in recent years, and three dimensional fabrication or “printing” systems have become an increasingly practical means of manufacturing organic and inorganic materials from a digital model. For clarity, three dimensional fabricators may be referred to as an additive manufacturing device or 3D Printer. A description of many such fabrication systems and recent developments in the art can be found in U.S. Pat. No. 7,625,198 to Lipson et al. and the patents and publications referenced therein.
  • With the proliferation of molecular biology research in recent years, the efficiency of researchers has proven to be an important factor in achieving successful results from research and development endeavors. With manual tasks such as micro-pipetting becoming a limiting factor on progress, a system to quicken and automate routine molecular biology procedures would have significant impacts on modern biology laboratories. A pipette (also called a pipet, pipettor, or chemical dropper) is a laboratory tool used to transport a measured volume of liquid. Pipettes are commonly used in molecular biology, analytical chemistry, and medical tests. Pipettes come in several designs for various purposes with differing levels of accuracy and precision, from single piece glass pipettes to more complex adjustable or electronic pipettes. Many pipette types work by creating a partial vacuum above the liquid-holding chamber and selectively releasing this vacuum to draw up and dispense liquid.
  • Though many experiments depend on cell geometry, to date, current activities molecular biology may be hindered by the inability to standardize cell culture across experiments, often relying on inaccurate and subjective measures of cell culture growth and confluence. This has thereby led to severe consequences because a basic tenet of scientific method is replication (i.e., to repeat an experiment in order to duplicate the results, thus further validating the underlying hypothesis), something which current techniques do not allow in precise terms.
  • In addition, and often in an effort to achieve more precise and reliable results, current pipetting activities may need to be performed by individuals with sufficient training and developed skills in pipetting procedures, which precludes laboratory assistants or students with less experience from performing experiments that require precision or reliability.
  • Due to the inherent complexities of performing experiments with identical samples and the shortcomings in currently known techniques, existing pipetting procedures may fail to achieve optimal levels of efficiency, reliability and accuracy. In particular, it would be desirable to have pipetting systems and techniques that increase the speed of conducting such procedures, minimize reliance on the individual skills of personnel involved in the experiments, and reduce the risk of generating flawed, unreliable or imprecise results.
  • SUMMARY OF THE INVENTION
  • The shortcomings of the prior art can be overcome and additional advantages can be provided with the Automated Visual Pipetting (“AVP”) systems and techniques described herein. The present invention can improve the accuracy, ease of use and efficiency of pipetting procedures in order to achieve a drastic improvement in performance and quality.
  • Embodiments of the present invention involving Automated Visual Pipetting are conceived with a few goals in mind, most notably: (1) to be visually consistent with the logic of experiments performed by hand; (2) such that a first day undergraduate research associate could be capable of operating such an embodiment, and running experiments within minutes of engaging it; and (3) able to be added to machines that may have non-pipetting uses, such as three dimensional fabricators (i.e., additive manufacturing) or other computer numerical control (CNC) machine tools like milling.
  • In one Automated Visual Pipetting embodiment of the present invention, a camera is mounted on a pipette or another type of tool that may be a deposition or non-deposition tool (e.g., milling). A camera feed and recognition software can enable users to replace existing micro-pipetting techniques with a computerized process that may be controlled with a few simple mouse clicks, while the user can directly visualize a live experimental setup. The embodiment also allows for integration with common molecular biology procedure “kits,” and enables the process to be automated and visualized without requiring constant interaction by the user.
  • Such embodiments of the present invention may be uniquely capable of pipetting and automating cell culture using a single machine and essentially the same software. Unlike existing pipetting techniques and systems that suffer the problems described above, embodiments of the present invention as disclosed herein can allow for cells to be deposited in exact (or very near exact) geometric patterns, providing several new opportunities such as optimizing repeatable cell line-specific deposition patterns for achieving desired confluence in a specific time frame, and experimenting using novel geometric cell arrangements which could potentially include platforms for studying single neuron synapses or creating entire “manufactured” organ systems.
  • Embodiments of the present invention may involve a camera and pipetting tool that, for example, is guided by a control unit receiving instructions from a fabrication and/or pipetting command unit (which may be a computer) running either a locally-stored or server-based fabrication and/or pipetting software application. The computer may be integrated into a fabrication and/or pipetting device, or connected to the fabrication and/or pipetting device via a wireless connection such as Bluetooth, WLAN, NFC or other wireless communication technologies, or a wired connection such as Ethernet, USB, FireWire, serial or parallel connection, or other wired communication technologies.
  • Some of the features provided by the system of the present disclosure are described as follows:
  • A three dimensional pipetting device, having a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head, and a plurality of interchangeable pipetting tips that can be affixed to the pipetting tool head, such that the control unit can operate the pipetting tool head to selectively use one of the plurality of interchangeable pipetting tips, and selectively draw up or dispense liquid from or to one or more containers on a work surface.
  • A three dimensional pipetting device, having a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head, and a camera for recording data including the position of one or more containers on a work surface, such that the control unit can operate the pipetting tool head to selectively draw up or dispense liquid from or to the one or more containers on the work surface. Additionally, the three dimensional pipetting device may also be a three dimensional fabricating system. Additionally, the pipetting command unit is configured to receive and process data recorded by the camera to determine the location and size of at least one container on the work surface. Additionally, the pipetting command unit is configured to receive and process data recorded by the camera to generate a digital image of a plurality of items on the work surface, which may optionally be output to an external monitor to display a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface. Additionally, the pipetting command unit is configured to receive data recorded by the camera and visually simulate the performance of a procedure that can be performed by the pipetting tool head. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera to identity an item on the work surface. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively draw up liquid from at least one container on the work surface. Additionally, the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively deposit liquid into at least one container on the work surface.
  • A method for using a three dimensional pipetting device, including the steps of transmitting instructions for operating a camera from a pipetting command unit to a control unit, operating the camera with the control unit to record image data of one or more containers on a work surface, transmitting recorded image data from the control unit to the pipetting command unit, transmitting instructions for operating a pipetting tool head from a pipetting command unit to a control unit; and operating the pipetting tool head to selectively draw up liquid from the one or more containers on the work surface. Additionally, the pipetting command unit processes recorded image data and determines the location of at least one container on the work surface. Additionally, the pipetting command unit processes recorded image data and generates a digital image of a plurality of items on the work surface, and the digital image may include a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface. Additionally, the pipetting command unit processes recorded image data received from the control unit and simulates performance of a procedure that can be performed by the pipetting tool head. Additionally, the pipetting command unit processes recorded image data and detects a visual indicator (that may be a QR code) in at least one of the recorded images that identifies an item on the work surface. Additionally, the pipetting command unit processes recorded image data, detects a visual indicator (that may be a QR code) in at least one of the recorded images as an instruction to use the pipetting tool head, and selectively draws up or deposits liquid from or into at least one container on the work surface in response to said instruction. Additionally, the pipetting command unit processes recorded image data received from the control unit and detects whether a prescribed amount of liquid has been aspirated from the one or more containers on the work surface by the pipetting tool head.
  • One great benefit of the present invention is that many embodiments of Automated Visual Pipetting can be combined with, and implemented in, three dimensional fabricators (3D printers). However, although many embodiments of the present invention described herein relate to a joint three dimensional fabricator and Automated Visual Pipetting device, it should be readily apparent to the reader that stand-alone Automated Visual Pipetting devices are within the scope of the present invention, and that a three dimensional fabricator is not required to implement the present invention. The present invention has many embodiments, some of which are described herein, and others which should be apparent to the reader or inferred from what is taught herein.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a perspective view of a three dimensional fabricator.
  • FIG. 2 shows an embodiment of the invention in which a camera is affixed to the deposition tool head.
  • FIG. 3 shows an embodiment of the invention in which a visual code is used by the on-board camera to guide motion and interchange tips, or identify items on the machine's build-tray.
  • FIG. 4 is a perspective view of a three dimensional pipetting device in accordance with an embodiment of the invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT(S)
  • In order to provide some background regarding three dimensional fabricating systems and illustrate common components in such devices that may be used in connection with the present invention, FIG. 1 provides a perspective view of a prior art three dimensional fabricating system. Fabrication system 100 includes fabricator 101 with material deposition tool head 102 (also referred to herein as deposition tool or deposition head), control unit 103 having one or more actuators and sensors configured to control operating characteristics of material deposition tool 102, and build tray (i.e., build surface) 104. Fabrication command unit 105 may be coupled to fabricator 101 as a component physically inside fabricator 101, or it may be coupled as an external device (e.g., computer) via a wired or wireless connection.
  • Fabrication command unit 105 includes processor 106, memory 107, and fabrication software application 108 that can be stored in memory 107 and executed by processor 106. It should be appreciated that control unit 103 of fabricator 101 may be configured to receive instructions from fabrication command unit 105 such that fabricator 101 can fabricate an output product on build surface 104 from materials dispensed by material deposition tool 102. Fabrication software application 108 can generate tool path information for fabricator 101 and delineate how material can be used to generate shapes. Complex CAD programs may also be used to generate the intended geometry.
  • Embodiments of the present invention may be implemented in any three dimensional fabricating system (i.e., additive manufacturing device or 3D Printer), for example, as illustrated in FIG. 1 and described above, that is suitable for performing Automated Visual Pipetting techniques. Other exemplary three dimensional fabricating system or components thereof that may be suitable for Automated Visual Pipetting are described in U.S. Pub. No. 2012/0241993 entitled “SYSTEMS AND METHODS FOR FREEFORM FABRICATION OF FOAMED STRUCTURES” and published on Sept. 27, 2012 (filed as U.S. application Ser. No. 13/356,194 on Jan. 23, 2012) and U.S. Pat. No. 7,625,198 to Lipson et al.
  • By way of further explanation, embodiments of the present invention may use Fabrication system 100 to perform Automated Visual Pipetting. Material deposition tool 102 may include a mounted pipetting tool and camera, or be replaced with a pipetting tool and camera. Alternatively, material deposition tool 102 may include a mounted pipetting tool or be replaced with a pipetting tool, and one or more cameras may be mounted on other surfaces of fabricator 101. Control unit 103 having one or more actuators and sensors configured to control operating characteristics of material deposition tool 102, may similarly be configured to control operating characteristics of a pipetting tool and one or more cameras. Experiments to undergo pipetting procedures may be placed onto build tray (i.e., build surface) 104.
  • Fabrication command unit 105 may be configured to support manual and automated use of a pipetting tool and one or more cameras. Likewise, fabrication software application 108 may include or be replaced with Automated Visual Pipetting software to receive pipetting and/or camera requests from a user, generate pipetting tool and/or camera path information, and direct fabricator 101 and perform manual and automated pipetting procedures, as well as operate the one or more cameras.
  • Automated Visual Pipetting software included with, or replacing, fabrication software application 108 can be stored in memory 107 and executed by processor 106. It should be appreciated that control unit 103 of fabricator 101 may be configured to receive instructions from fabrication command unit 105 such that fabricator 101 can conduct pipetting procedures on experiments placed upon build surface 104 from a pipetting tool mounted on, or replacing, deposition tool 102. Therefore, in various embodiments of the present invention, fabrication system 100 may be referred to as AVP system 100, fabrication command unit 105 may be referred to as AVP command unit 105, and fabricator 101 may be referred to as a three dimensional pipetting device, as it supports three dimensional movement of the pipetting tool above build surface 104.
  • Embodiments of the present invention may be implemented in three dimensional fabricating systems (i.e., additive manufacturing devices or 3D Printers), similar to the example illustrated in FIG. 1 and described above. U.S. Pub. No. 2012/0241993 entitled “SYSTEMS AND METHODS FOR FREEFORM FABRICATION OF FOAMED STRUCTURES” and published on Sep. 27, 2012 (filed as U.S. application Ser. No. 13/356,194 on Jan. 23, 2012) provides a similar depiction of FIG. 1 with an accompanying disclosure. As noted above, additional information regarding many components of three dimensional fabricating systems may be found in U.S. Pat. No. 7,625,198 to Lipson et al.
  • As shown in an embodiment of the present invention depicted in FIG. 2, tool head 201 can accept a plurality of disposable (or reusable) pipette tips (or needles) 202 and act as a pipette or micropipette would function in the hand of a biologist. Camera unit 203 can send a live video or image feed of lens 204 to a computer (e.g., Fabrication command unit 105 of FIG. 1). The live video or image feed can be used by the computer to map out an experimental setup placed on a build surface (e.g., build tray 104 of FIG. 1) which is accessible to tool head 201. Tool head 201 may be mounted to, or replace, material deposition tool 102 of FIG. 1.
  • By using the camera, Automated Visual Pipetting software (e.g., included with, or replacing, fabrication software application 108) can assess where items are in a setup by visualizing items, the particular containers being used, or by other visual indicators, such as a particular shape, a QR code, a bar code, one or more numbers and/or letters, a specific packaging design or logo, etc. Camera 203 may be used by AVP system 100 to automatically guide tool head 201, or to assess distance, for example by measuring pixels between two points of known distance and thereby calculating a distance from those points by how big they appear to the camera (on a pixel basis). Camera 203 may be able to visualize multiple resolutions, by the use of digital or optical zoom. This could be used to integrate information about experimental results in real time to the software. Camera 203 could be calibrated to detect different colored circumferences on various types of containers, to easily identify components of a commonly used kit of experimental materials. Additionally, by using the camera, Automated Visual Pipetting software (e.g., included with, or replacing, fabrication software application 108) can assess what pipetting operations to conduct with respect to particular items detected by the camera (e.g., drawing liquids or depositing liquids) by visualizing instructions on, e.g., a container or build tray, such as by recognizing the particular containers being used, or by other visual indicators, such as a particular shape, a QR code, a bar code, one or more numbers and/or letters, a specific packaging design or logo, etc. The visual indicator may directly identify the procedure to be performed; which, with the use of a kit, may require the user to identify a substrate on which to perform the procedure. Automated Visual Pipetting software may read instructions directly from the items, or suggest (and execute) possible procedures based on the identity of items on the tray and their corresponding potential uses in a series of known procedures.
  • In certain embodiments, Camera 203 may allow a user viewing a live video or image feed from Camera 203 via the Automated Visual Pipetting software to guide tool head 201 with, e.g., computer commands, a mouse or a joystick. However, the Automated Visual Pipetting software may be designed to limit the actions of a user in order to prevent errors or unintended consequences resulting from user mistakes. For example, the software may prevent the user from initiating a pipetting action when tool head 201 is not aligned properly over the target to be pipetted. As another example, the software may prevent the user from initiating a pipetting action when tool head 201 is aligned over a sample that should not be pipetted (or has already been pipetted).
  • Embodiments of the present invention may cover the entire spectrum of AVP functionality, with embodiments on one end of the spectrum completely automating the described procedures, and embodiments at the other end allowing these procedures to be manually directed by a user of the computer running the Automated Visual Pipetting software. Different implementations involving partial automation and partial user interaction will be readily apparent to one of ordinary skill in the art and constitute embodiments of the present invention as well. A single device may be configurable such that it can implement a number of different embodiments, each one involving different levels of automation and user interaction.
  • FIG. 3 illustrates additional aspects of an embodiment of the present invention. Atop build table 301 (e.g., build tray 104 of FIG. 1), a plurality of kits and accessories can be placed, such as the standard pipette tip box 305. Tip box 305 has several holes 306 in which pipette tips 309 are removed by tool head 201 of FIG. 2 (as illustrated at the top of FIG. 3). A camera (e.g., Camera 203 of FIG. 2) can identify target 308 on box 305. By knowing the actual size of target 308, AVP system 100 can determine the distance of tool head 201 from box 305 by the relative (pixel) size of target 308. Alternatively, AVP system 100 can measure the distance in pixels between target rings 303 and 304 to determine the distance between tool head 201 and box 305 or build table 301. Having target rings 303 and 304 on build table 301, is one possible way to eliminate the need for target 308 on box 305, since it has a feature or object of known dimension. Alternatively, QR code 307 or QR code 302 can be used to identify a particular kit (e.g., box 305) or build table 301, respectively. Just as the distance can be calculated using target 308, it can also be calculated simply by using the relative (pixel) size of a pipette tip's visible, proximal lumen and comparing it to the known size of a pipette tip's visible lumen in a tip box. The same technique can be used to determine the distance from any other object, such as a conical tube, which could be visually matched by the Automated Visual Pipetting software to a likely, known tube and then distance measured by the relative pixel measuring system just described. Even if the tube were not directly in line below a camera, Automated Visual Pipetting software can be adjusted to compensate by noting the difference in appearance of a tube that was perfectly flat. Likewise, a camera could also be used to detect the positioning of material inside a vessel (e.g. conical tube), and then position the pipette tip to capture that material, and also measure the volume of material aspirated (e.g. by visualizing the extent to which the pipette tip has been filled) to assure it was successful. (Such visualization may require the use or a secondary camera or mirror system to the primary camera to visualize the engaged deposition pipette tip.) Often, if there is insufficient liquid to aspirate the amount of fluid prescribed by the user of a micropipette system, as the present invention, a jet of aspirated air will cause the liquid to splash inside the pipette tip. Likewise, the knowledge of the fluid properties (e.g. viscosity, surface tension, density, etc.) allows the present system to predict the level at which the properly aspirated amount of fluid should rise to. By visually examining the tip, through direct camera visualization or by way of a mirror angled to allow camera visualization, the present invention can discern whether a uniform aspiration was carried out, (e.g. was there a jet of air caused by lack of material) or whether liquid rises to the proper level in the tip (e.g. as in cases both where there was or was not an air jet, as might be the case for a viscous fluid). The software may use a comparative algorithm between an empty tip and the aspirated state (or expectations thereof) to perform the above functionality.
  • Laboratory procedures using an embodiment of the present invention may be performed in a variety of ways: by using the live video or image feed, tool head 201 could be controlled by simple mouse clicks (or other input methods) on a connected computing device, which may be connected directly, or over a local wireless connection, or via the Internet. These techniques could replace the need for skilled pipetting procedures by providing similar results with as little as a few computer commands (e.g., a few mouse clicks), allowing a live experiment to be run by a user without special training or manual skills. Moreover, these techniques could be initiated with a virtual run through, and then performed without requiring any user interaction. For example, a user could make a few selections on a computer to initiate an experiment, the computer could visually present a virtual demonstration of that procedure, await for user confirmation to begin, and then perform the process without requiring further user interaction. Alternatively, a procedure can be programmed for a particular experimental setup. In this scenario, the experimenter would simply have to place the appropriate components, kits, and accessories on a build table, and AVP system 100 would recognize these items and simply proceed with performing an experiment based on a preprogrammed set of instructions. These instructions need not be relative to the position of items on the build table, but rather to what the items actually are, which is now possible due to the use of a video camera. AVP system 100 could identify the various components, or allow the user to do so, and then proceed with the experiment as would a person, being able to adapt to the “random” positioning of objects by knowing what and where things are, not just where they should be.
  • These embodiments could use interchangeable (optionally disposable and/or sterilizable) pipette (needle) tips, much in the same way that many of the deposition heads described above can switch tips, and though the mechanisms may differ slightly to accommodate mass produced pipette tips currently available, implementation of embodiments would be apparent in view of the disclosures herein.
  • These embodiments can also automate the use of cell cultures by processing cell growth data and depositing cells in geometries which will lead to a particular result, such as a desired cell confluence (in x days), or yield a particular growth pattern, like a row of neurons.
  • For example, the embodiments can run as follows. Several items can be placed on a build tray, they will each be digitally identified either by reference to a database of known items, by a scanned code, or by object properties (e.g., opening size, position), all of which can use digital mapping via the camera. Each item can then be superimposed with a digital outline shape, showing the computer's recognition of an opening, and either identifying what an object is or allowing the object's identity to be assigned. With objects known, a pre-programmed protocol can be performed. Alternatively, placing the items on a tray can be a pre-defined indicator for what protocol the computer should run. If no protocol exists, a series of selections on a computer (i.e., with Automated Visual Pipetting software) can be made to indicate the pipetting action to be taken. This can be performed either in real time or be input into a computer, optionally previewed in a virtual run through, and then performed AVP system 100 will take appropriate measurements to assure it pipettes correct volume, as well as use the camera to help identify optimal pipette placement in a container and the amount of fluid in it (recall the z-axis is virtually a constant).
  • Embodiments can also involve creating a custom build tray (e.g., using a plastic deposition tool) and then setting up a unique experimental setup in that tray. After use, the tray can be discarded, replaced, cleaned or reused.
  • Embodiments could take a crowded build (work) tray with many tube trays and other components on it, and then visually separate those components out when displaying them on a computer screen (for example by identifying component edges or small areas of unused space), so that they are easier to visualize for the user, and not crowded in a displayed video feed despite their physical positioning. This can allow more efficient utility of build tray space, and potentially allow for more logical click-through protocol setup. A user could, for example, digitally rearrange the position of components as they appear on the computer screen (versus how they actually sit on the tray), and customize the visualization describing data about those components on the same screen.
  • Embodiments could also track the contents of each individual component over a series of experiments. So, for example, a QR code (or otherwise visually) labeled 12-well plate could have reagent X added to it on day one, incubated overnight, and then upon replacement onto the build (work) tray on day two, the camera would check the plate data against a database containing recorded actions of AVP system 100 on the previous day, such that the user would be aware that reagent X was added to particular wells on the plate the previous day. A series of tubes could have labels in several areas, for example the cap. In such a case, the user may wish to leave the closed tubes on the tray, allow the camera to identify the codes, and remove and replace one tube at a time (to reduce error) and then open them. Embodiments can also contain additional cameras at various positions (and/or with various angles) on one or more tool heads or on the device frame itself for a variety of advances uses, including reading identification codes not easily identified from above the build (work) tray.
  • Embodiments of the present invention could also be used with an integrated label gun, which would add labels to components in an experiment as needed.
  • By using a camera feed in certain embodiments, automation and manufacturing tasks can be more precisely controlled and audited. For example, a camera can be used to verify the quality of three dimensional items as they are being printed, or precisely pick and place, or pipette, deposition tasks.
  • FIG. 4 illustrates one possible embodiment with a modified version of Fabrication system 100 as shown in FIG. 1 and described herein. In this embodiment, AVP system 400 can perform Automated Visual Pipetting. Pipetting tool and camera 409 (such as the pipetting tool and camera illustrated in FIG. 2 and described herein) is mounted for use on deposition tool head 402. Control unit 403 has one or more actuators and sensors configured to control operating characteristics of pipetting tool and camera 40×. Experiments to undergo pipetting procedures may be placed onto work tray (i.e., work surface) 404. For example, Container 410 may be a beaker or petri dish and contain an item such as a liquid, substrate or cell culture. Pipette tip box 411 (such as the pipette tip box illustrated in FIG. 3 and described herein) may also be placed on work tray 404 within reach of pipetting tool and camera 409.
  • AVP command unit 405 may be configured to support manual and automated use of pipetting tool and camera 409. Likewise, AVP software application 408 may receive pipetting and/or camera requests from a user, generate pipetting tool and/or camera path information, and direct control unit 403 to perform manual and automated pipetting procedures, as well as operate one or more cameras.
  • Automated Visual Pipetting software included can be stored in memory 407 and executed by processor 406. It should be appreciated that control unit 403 of AVP system 400 may be configured to receive instructions from AVP command unit 405 such that AVP system 400 can conduct pipetting procedures on experiments placed upon work surface 404 from pipetting tool and camera 409 mounted on deposition tool 402. Therefore, in various embodiments of the present invention, AVP system 400 may be referred to as a three dimensional pipetting device because it supports two axis movement of the pipetting tool above work surface 404, as well as up and down movement.
  • AVP command unit 405 may have a wireless or wired connection to external computer screen (i.e., monitor) 413 direct output of recorded images from pipetting tool and camera 409 to external computer screen 413, such as to display image 415 of container 410, and image 416 of pipette tip box 411. As discussed above, a computer screen can also be used in a variety of way to take advantage of the benefits offered by embodiments of the present invention.
  • It will be appreciated by persons of ordinary skill in the art that the present invention is not limited to the exemplary embodiments illustrated and described herein, nor is it limited to the dimensions or specific physical implementations illustrated and described herein. The present invention may have other embodiments that are readily apparent and enabled as a result of the concepts and descriptions provided herein.

Claims (20)

What is claimed is:
1. A three dimensional pipetting device, comprising:
a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head; and
a plurality of interchangeable pipetting tips that can be affixed to the pipetting tool head;
wherein the control unit can operate the pipetting tool head to selectively use one of the plurality of interchangeable pipetting tips; and
wherein the control unit can operate the pipetting tool head to selectively draw up liquid from one or more containers on a work surface;
wherein the control unit can operate the pipetting tool head to selectively dispense liquid into the one or more containers on the work surface.
2. A three dimensional pipetting device, comprising:
a control unit for receiving instructions from a pipetting command unit and operating a pipetting tool head; and
a camera for recording data including the position of one or more containers on a work surface;
wherein the control unit can operate the pipetting tool head to selectively draw up liquid from the one or more containers on the work surface; and
wherein the control unit can operate the pipetting tool head to selectively dispense liquids into the one or more containers on the work surface.
3. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive and process data recorded by the camera to determine the location and size of at least one container on the work surface.
4. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive and process data recorded by the camera to generate a digital image of a plurality of items on the work surface.
5. A three dimensional pipetting device of claim 4, wherein the digital image of the plurality of items on the work surface can be output to an external monitor to display a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface.
6. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive data recorded by the camera and visually simulate the performance of a procedure that can be performed by the pipetting tool head.
7. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive and process a visual indicator recorded by the camera to identity an item on the work surface.
8. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively draw up liquid from at least one container on the work surface.
9. A three dimensional pipetting device of claim 2, wherein the pipetting command unit is configured to receive and process a visual indicator recorded by the camera as an instruction to use the pipetting tool head and selectively deposit liquid into at least one container on the work surface.
10. A three dimensional pipetting device of claim 2, wherein the three dimensional pipetting device is also a three dimensional fabricating system.
11. A method for using a three dimensional pipetting device, comprising the steps of:
transmitting instructions for operating a camera from a pipetting command unit to a control unit;
operating the camera with the control unit to record image data of one or more containers on a work surface;
transmitting recorded image data from the control unit to the pipetting command unit;
transmitting instructions for operating a pipetting tool head from a pipetting command unit to a control unit; and
operating the pipetting tool head to selectively draw up liquid from the one or more containers on the work surface.
12. The method of claim 11, further comprising the step of the pipetting command unit processing recorded image data received from the control unit and determining the location of at least one container on the work surface.
13. The method of claim 11, further comprising the step of the pipetting command unit processing recorded image data received from the control unit and generating a digital image of a plurality of items on the work surface.
14. The method of claim 13, wherein the digital image includes a virtual arrangement of the items that is different from the physical arrangement of the items on the work surface.
15. The method of claim 11, wherein after said step of transmitting recorded image data from the control unit to the pipetting command unit, and before said step of transmitting instructions for operating a pipetting tool head from a pipetting command unit to a control unit, the pipetting command unit processes recorded image data received from the control unit and simulates performance of a procedure that can be performed by the pipetting tool head.
16. The method of claim 11, further comprising the steps of:
the pipetting command unit processing recorded image data received from the control unit;
detecting a visual indicator in at least one of the recorded images; and
identifying an item on the work surface.
17. The method of claim 16, wherein said visual indicator is a QR code.
18. The method of claim 11, further comprising the steps of:
the pipetting command unit processing recorded image data received from the control unit;
detecting a visual indicator in at least one of the recorded images as an instruction to use the pipetting tool head; and
selectively drawing up liquid from at least one container on the work surface in response to said instruction.
19. The method of claim 11, further comprising the steps of:
the pipetting command unit processing recorded image data received from the control unit;
detecting a visual indicator in at least one of the recorded images as an instruction to use the pipetting tool head; and
selectively depositing liquid into at least one container on the work surface in response to said instruction.
20. The method of claim 11, further comprising the steps of:
the pipetting command unit processing recorded image data received from the control unit; and
detecting whether a prescribed amount of liquid has been aspirated from the one or more containers on the work surface by the pipetting tool head.
US13/763,715 2012-02-10 2013-02-10 Automated visual pipetting Abandoned US20130205920A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/763,715 US20130205920A1 (en) 2012-02-10 2013-02-10 Automated visual pipetting

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201261633433P 2012-02-10 2012-02-10
US201261689963P 2012-06-18 2012-06-18
US201261741368P 2012-07-18 2012-07-18
US13/763,715 US20130205920A1 (en) 2012-02-10 2013-02-10 Automated visual pipetting

Publications (1)

Publication Number Publication Date
US20130205920A1 true US20130205920A1 (en) 2013-08-15

Family

ID=48944526

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/761,272 Expired - Fee Related US9884318B2 (en) 2012-02-10 2013-02-07 Multi-axis, multi-purpose robotics automation and quality adaptive additive manufacturing
US13/763,715 Abandoned US20130205920A1 (en) 2012-02-10 2013-02-10 Automated visual pipetting

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/761,272 Expired - Fee Related US9884318B2 (en) 2012-02-10 2013-02-07 Multi-axis, multi-purpose robotics automation and quality adaptive additive manufacturing

Country Status (1)

Country Link
US (2) US9884318B2 (en)

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103522547A (en) * 2013-09-26 2014-01-22 上海大学 Numerically-controlled machine tool power-driven 3D (three dimensional) printing head component and method for manufacturing three-dimensional support
CN104908321A (en) * 2015-05-22 2015-09-16 杭州汉卓机电科技有限公司 Control method of pneumatic double-acting biological 3D printing nozzle with internal and external double spiral cover structures
WO2015172025A1 (en) * 2014-05-08 2015-11-12 The Cleveland Clinic Foundation Systems and methods for detection, analysis, isolation and/or harvesting of biological objects
US20150323920A1 (en) * 2013-01-18 2015-11-12 Trumpf Werkzeugmaschinen Gmbh + Co. Kg Machine Maintenance Using a Machine Controller and a Service Computer
CN105216312A (en) * 2015-10-12 2016-01-06 周末 A kind of 3D printing device checkout gear
US20160011083A1 (en) * 2014-07-11 2016-01-14 Intellicyt Contact Sensing Probe and Methods of Use for Microplate Liquid Sampling
CN105705319A (en) * 2013-09-19 2016-06-22 马克弗巨德有限公司 Methods for fiber reinforced additive manufacturing
CN106292416A (en) * 2015-06-24 2017-01-04 英国电讯有限公司 The gate printed
WO2017001676A1 (en) * 2015-07-02 2017-01-05 Hamilton Bonaduz Ag Pipetting apparatus with image processing
CN107000318A (en) * 2014-12-01 2017-08-01 沙特基础工业全球技术有限公司 The nozzle instrument for extruding increasing material manufacturing for material changes
WO2017157402A1 (en) * 2016-03-17 2017-09-21 Itu Business Development A/S A robot and a method of controlling a robot
US9815268B2 (en) 2013-03-22 2017-11-14 Markforged, Inc. Multiaxis fiber reinforcement for 3D printing
US9956725B2 (en) 2013-03-22 2018-05-01 Markforged, Inc. Three dimensional printer for fiber reinforced composite filament fabrication
US10016942B2 (en) 2013-03-22 2018-07-10 Markforged, Inc. Three dimensional printing
US10040252B2 (en) 2013-03-22 2018-08-07 Markforged, Inc. Methods for fiber reinforced additive manufacturing
US10076875B2 (en) 2013-03-22 2018-09-18 Markforged, Inc. Methods for composite filament fabrication in three dimensional printing
US10076876B2 (en) 2013-03-22 2018-09-18 Markforged, Inc. Three dimensional printing
US10099427B2 (en) 2013-03-22 2018-10-16 Markforged, Inc. Three dimensional printer with composite filament fabrication
US20190054682A1 (en) * 2016-05-12 2019-02-21 Hewlett-Packard Development Company, L.P. Build material splash control
US10259160B2 (en) 2013-03-22 2019-04-16 Markforged, Inc. Wear resistance in 3D printing of composites
US10603841B2 (en) 2013-03-22 2020-03-31 Markforged, Inc. Multilayer fiber reinforcement design for 3D printing
US10611082B2 (en) 2013-03-22 2020-04-07 Markforged, Inc. Apparatus for fiber reinforced additive manufacturing
US10682844B2 (en) 2013-03-22 2020-06-16 Markforged, Inc. Embedding 3D printed fiber reinforcement in molded articles
US20200408791A1 (en) * 2019-06-28 2020-12-31 Sartorius Biohit Liquid Handling Oy Method of information transmission, a liquid handling device, and a system
US10953609B1 (en) 2013-03-22 2021-03-23 Markforged, Inc. Scanning print bed and part height in 3D printing
US20210181222A1 (en) * 2018-04-23 2021-06-17 Shimadzu Corporation Autosampler
US11065861B2 (en) 2013-03-22 2021-07-20 Markforged, Inc. Methods for composite filament threading in three dimensional printing
US20210299870A1 (en) * 2018-08-24 2021-09-30 The University Of Tokyo Robot assistance device and robot assistance system
US11135579B2 (en) 2016-07-13 2021-10-05 Hewlett-Packard Development Company, L.P. Apparatus with encoded media to indicate dispensing locations for pipette dispenser
US11148409B2 (en) 2013-03-22 2021-10-19 Markforged, Inc. Three dimensional printing of composite reinforced structures
WO2022013141A1 (en) * 2020-07-16 2022-01-20 Signify Holding B.V. Single camera-based process monitoring in fdm 3d printer
US11237542B2 (en) 2013-03-22 2022-02-01 Markforged, Inc. Composite filament 3D printing using complementary reinforcement formations
US11263433B2 (en) 2016-10-28 2022-03-01 Beckman Coulter, Inc. Substance preparation evaluation system
JP2022518066A (en) * 2019-01-28 2022-03-11 フォーミュラトリックス・インコーポレイテッド Liquid processing equipment and pipette heads and methods for sucking and / or distributing liquids
CN116728778A (en) * 2023-07-17 2023-09-12 北京科技大学 Basalt fiber composite material 3D printing process equipment
US11787104B2 (en) 2013-03-22 2023-10-17 Markforged, Inc. Methods for fiber reinforced additive manufacturing
US11981069B2 (en) 2013-03-22 2024-05-14 Markforged, Inc. Three dimensional printing of composite reinforced structures

Families Citing this family (106)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9511543B2 (en) 2012-08-29 2016-12-06 Cc3D Llc Method and apparatus for continuous composite three-dimensional printing
US9102098B2 (en) 2012-12-05 2015-08-11 Wobbleworks, Inc. Hand-held three-dimensional drawing device
US20140232035A1 (en) * 2013-02-19 2014-08-21 Hemant Bheda Reinforced fused-deposition modeling
US9643362B2 (en) * 2013-03-15 2017-05-09 Microsoft Technology Licensing, Llc Full color three-dimensional object fabrication
US9126367B1 (en) 2013-03-22 2015-09-08 Markforged, Inc. Three dimensional printer for fiber reinforced composite filament fabrication
US9539762B2 (en) 2013-03-22 2017-01-10 Markforged, Inc. 3D printing with kinematic coupling
US10543549B2 (en) * 2013-07-16 2020-01-28 Illinois Tool Works Inc. Additive manufacturing system for joining and surface overlay
GB201314030D0 (en) * 2013-08-06 2013-09-18 Eads Uk Ltd Extrusion-Based Additive Manufacturing System and Method
US11077607B2 (en) * 2013-10-21 2021-08-03 Made In Space, Inc. Manufacturing in microgravity and varying external force environments
US10052820B2 (en) 2013-09-13 2018-08-21 Made In Space, Inc. Additive manufacturing of extended structures
CN103521765B (en) * 2013-09-29 2016-04-20 福建海源三维打印高科技有限公司 A kind of Portable 3D printer
US10086568B2 (en) 2013-10-21 2018-10-02 Made In Space, Inc. Seamless scanning and production devices and methods
US10725451B2 (en) 2013-10-21 2020-07-28 Made In Space, Inc. Terrestrial and space-based manufacturing systems
US9085109B2 (en) 2013-11-15 2015-07-21 Makerbot Industries, Llc Three-dimensional printer tool systems
WO2015073301A1 (en) 2013-11-18 2015-05-21 Chang Kai-Jui Color or multi-material three-dimensional (3d) printing
US10436983B2 (en) * 2013-11-19 2019-10-08 Verizon Patent And Licensing Inc. Machine-to-machine plant automation using 3D printed fiber splicing
US20150197063A1 (en) * 2014-01-12 2015-07-16 Zohar SHINAR Device, method, and system of three-dimensional printing
US10611098B2 (en) 2014-01-17 2020-04-07 G6 Materials Corp. Fused filament fabrication using multi-segment filament
US10052797B2 (en) 2014-01-25 2018-08-21 Made In Space, Inc. Recycling materials in various environments including reduced gravity environments
US10307970B2 (en) 2014-02-20 2019-06-04 Made In Space, Inc. In-situ resource preparation and utilization methods
CA2884018C (en) 2014-02-26 2022-06-21 Freespace Composites Inc. Manufacturing system using topology optimization design software, novel three-dimensional printing mechanisms and structural composite materials
CN103802238A (en) * 2014-02-26 2014-05-21 英华达(上海)科技有限公司 Single-nozzle color feeding method and device for 3D (Three-Dimensional) printer
CN103847101A (en) * 2014-02-26 2014-06-11 珠海天威飞马打印耗材有限公司 Three-dimensional printer
DE102014207507B4 (en) 2014-04-17 2021-12-16 Kennametal Inc. Cutting tool and method for producing a cutting tool
DE102014207510B4 (en) 2014-04-17 2021-12-16 Kennametal Inc. Cutting tool and method for producing a cutting tool
TWI491496B (en) * 2014-05-08 2015-07-11 三緯國際立體列印科技股份有限公司 Three dimensional printing apparatus, printing calibration board and three dimensional printing calibration method thereof
JP6485617B2 (en) * 2014-06-20 2019-03-20 泰 金田 Three-dimensional printing method for arranging filaments horizontally without support
US20150375451A1 (en) * 2014-06-30 2015-12-31 Disney Enterprises, Inc. Multi-color 3d printer
US10737480B2 (en) * 2014-07-08 2020-08-11 Amril Ag Sinterable feedstock for use in 3D printing devices
KR20170038001A (en) 2014-07-22 2017-04-05 스트래터시스,인코포레이티드 Gear-based liquefier assembly for additive manufacturing system, and methods of use thereof
US9808991B2 (en) 2014-07-29 2017-11-07 Cc3D Llc. Method and apparatus for additive mechanical growth of tubular structures
CN104129079B (en) * 2014-08-07 2016-11-09 北京汇天威科技有限公司 A kind of novel 3D printer
US10870268B2 (en) 2014-08-21 2020-12-22 Mosaic Manufacturing Ltd. Series enabled multi-material extrusion technology
JP5969562B2 (en) * 2014-09-16 2016-08-17 株式会社東芝 Nozzle device and additive manufacturing device
US9643282B2 (en) 2014-10-17 2017-05-09 Kennametal Inc. Micro end mill and method of manufacturing same
WO2016081499A1 (en) * 2014-11-17 2016-05-26 Markforged, Inc. Composite filament 3d printing using complementary reinforcement formations
US10040235B2 (en) 2014-12-30 2018-08-07 Wobbleworks, Inc. Extrusion device for three-dimensional drawing
JP6547327B2 (en) * 2015-02-19 2019-07-24 セイコーエプソン株式会社 Three-dimensional object formation device, control method for three-dimensional object formation device, and control program for three-dimensional object formation device
KR102307599B1 (en) * 2015-03-16 2021-10-05 엘지전자 주식회사 Printing apparatus for building three-dimensional object
US9707721B2 (en) * 2015-04-03 2017-07-18 Thermwood Corporation Machine for forming thermoplastic articles and device utilized in the production thereof
CN106257979A (en) 2015-04-17 2016-12-28 沃博沃克斯有限公司 The distribution driving pressure of the peripheral of silk thread in extrusion device
DE102015207254A1 (en) 2015-04-21 2016-12-01 Eos Gmbh Electro Optical Systems Device and method for the generative production of a three-dimensional object
EP3106284A1 (en) * 2015-06-18 2016-12-21 Siemens Aktiengesellschaft Method and device for applying at least one material, extruder, 3d print head, 3d printer, machine tool and control device
EP3331659A4 (en) 2015-08-03 2019-08-14 Made In Space, Inc. In-space manufacturing and assembly of spacecraft device and techniques
US11170514B2 (en) * 2015-10-27 2021-11-09 Canon Kabushiki Kaisha Image processing apparatus, image processing method, 3D printing system, and storage medium
US10137679B2 (en) 2015-10-29 2018-11-27 Raytheon Company Material deposition system for additive manufacturing
US10471697B2 (en) * 2015-11-13 2019-11-12 R3 Printing, Inc. System and method for on-demand colorization for extrusion-based additive construction
CN105608275B (en) * 2015-12-23 2018-08-14 南京航空航天大学 A kind of generation method of printing path for model lightweight 3 D-printing
WO2017123726A1 (en) * 2016-01-12 2017-07-20 Markforged, Inc. Embedding 3d printed fiber reinforcement in molded articles
WO2017124085A1 (en) * 2016-01-15 2017-07-20 Markforged, Inc. Continuous and random reinforcement in a 3d printed part
US10105910B2 (en) 2016-04-15 2018-10-23 Cc3D Llc Method for continuously manufacturing composite hollow structure
US10232551B2 (en) 2016-04-15 2019-03-19 Cc3D Llc Head and system for continuously manufacturing composite hollow structure
WO2017194156A1 (en) 2016-05-12 2017-11-16 Hewlett-Packard Development Company L.P. Outlet structure
US10300659B2 (en) * 2016-06-23 2019-05-28 Raytheon Company Material deposition system for additive manufacturing
US10625467B2 (en) 2016-09-06 2020-04-21 Continuous Composites Inc. Additive manufacturing system having adjustable curing
US20180065317A1 (en) 2016-09-06 2018-03-08 Cc3D Llc Additive manufacturing system having in-situ fiber splicing
US10759113B2 (en) 2016-09-06 2020-09-01 Continuous Composites Inc. Additive manufacturing system having trailing cure mechanism
US11029658B2 (en) 2016-09-06 2021-06-08 Continuous Composites Inc. Systems and methods for controlling additive manufacturing
US10543640B2 (en) 2016-09-06 2020-01-28 Continuous Composites Inc. Additive manufacturing system having in-head fiber teasing
CN109982826B (en) * 2016-09-22 2021-11-09 Lg电子株式会社 Hydrogel discharge device
KR102478282B1 (en) * 2016-09-22 2022-12-15 엘지전자 주식회사 Hydrogel Discharge Apparatus and method for the same
US10766594B2 (en) 2016-11-03 2020-09-08 Continuous Composites Inc. Composite vehicle body
US20210094230A9 (en) 2016-11-04 2021-04-01 Continuous Composites Inc. System for additive manufacturing
US10953598B2 (en) 2016-11-04 2021-03-23 Continuous Composites Inc. Additive manufacturing system having vibrating nozzle
CN106665728A (en) * 2016-11-15 2017-05-17 贵州师范大学 3D-printing cake machine capable of automatically changing nozzles
US10857726B2 (en) 2017-01-24 2020-12-08 Continuous Composites Inc. Additive manufacturing system implementing anchor curing
US10040240B1 (en) 2017-01-24 2018-08-07 Cc3D Llc Additive manufacturing system having fiber-cutting mechanism
US10838401B2 (en) 2017-01-25 2020-11-17 Hewlett-Packard Development Company, L.P. Producing instructions that control three-dimensional printing from voxels
US20180229092A1 (en) 2017-02-13 2018-08-16 Cc3D Llc Composite sporting equipment
US10798783B2 (en) 2017-02-15 2020-10-06 Continuous Composites Inc. Additively manufactured composite heater
JP7365239B2 (en) * 2017-05-25 2023-10-19 ティーディービーティー アイピー インコーポレイティド Sterile printer system with dual arm mechanism
US10906240B2 (en) 2017-06-29 2021-02-02 Continuous Composites Inc. Print head for additive manufacturing system
US10814569B2 (en) 2017-06-29 2020-10-27 Continuous Composites Inc. Method and material for additive manufacturing
US10836108B1 (en) 2017-06-30 2020-11-17 Made In Space, Inc. System and method for monitoring and inspection of feedstock material for direct feedback into a deposition process
WO2019017929A1 (en) * 2017-07-19 2019-01-24 Hewlett-Packard Development Company, L.P. Three-dimensional printer
US10319499B1 (en) 2017-11-30 2019-06-11 Cc3D Llc System and method for additively manufacturing composite wiring harness
US10131088B1 (en) 2017-12-19 2018-11-20 Cc3D Llc Additive manufacturing method for discharging interlocking continuous reinforcement
CN108312541B (en) * 2017-12-28 2021-05-25 网云(武汉)三维科技股份有限公司 3D printer with rotatory structure of independently coloring
US10081129B1 (en) 2017-12-29 2018-09-25 Cc3D Llc Additive manufacturing system implementing hardener pre-impregnation
US10857729B2 (en) 2017-12-29 2020-12-08 Continuous Composites Inc. System and method for additively manufacturing functional elements into existing components
US11167495B2 (en) 2017-12-29 2021-11-09 Continuous Composites Inc. System and method for additively manufacturing functional elements into existing components
US10759114B2 (en) 2017-12-29 2020-09-01 Continuous Composites Inc. System and print head for continuously manufacturing composite structure
US10919222B2 (en) 2017-12-29 2021-02-16 Continuous Composites Inc. System and method for additively manufacturing functional elements into existing components
CH714727B1 (en) * 2018-03-07 2022-07-15 Coobx Ag 3D printer disposable cartridge.
US11161300B2 (en) 2018-04-11 2021-11-02 Continuous Composites Inc. System and print head for additive manufacturing system
US11110654B2 (en) 2018-04-12 2021-09-07 Continuous Composites Inc. System and print head for continuously manufacturing composite structure
US11110656B2 (en) 2018-04-12 2021-09-07 Continuous Composites Inc. System for continuously manufacturing composite structure
US10520923B2 (en) 2018-05-22 2019-12-31 Mantle Inc. Method and system for automated toolpath generation
CN108790157B (en) * 2018-05-25 2020-05-19 河北工业大学 Double-nozzle rapid forming system for environment-sensitive functionally-graded material
US11052603B2 (en) 2018-06-07 2021-07-06 Continuous Composites Inc. Additive manufacturing system having stowable cutting mechanism
US11292191B2 (en) 2018-06-15 2022-04-05 Rosemount Aerospace Inc. Multi-material fabrication with direct-write additive manufacturing
US20200086563A1 (en) 2018-09-13 2020-03-19 Cc3D Llc System and head for continuously manufacturing composite structure
US11235522B2 (en) 2018-10-04 2022-02-01 Continuous Composites Inc. System for additively manufacturing composite structures
US11511480B2 (en) 2018-10-26 2022-11-29 Continuous Composites Inc. System for additive manufacturing
US11420390B2 (en) 2018-11-19 2022-08-23 Continuous Composites Inc. System for additively manufacturing composite structure
US11358331B2 (en) 2018-11-19 2022-06-14 Continuous Composites Inc. System and head for continuously manufacturing composite structure
US20200238603A1 (en) 2019-01-25 2020-07-30 Continuous Composites Inc. System for additively manufacturing composite structure
US11312083B2 (en) 2019-05-28 2022-04-26 Continuous Composites Inc. System for additively manufacturing composite structure
CN110978520A (en) * 2019-12-24 2020-04-10 芜湖市爱三迪电子科技有限公司 Thickness adjusting device is used in processing of 3D printing consumables
US11840022B2 (en) 2019-12-30 2023-12-12 Continuous Composites Inc. System and method for additive manufacturing
US11904534B2 (en) 2020-02-25 2024-02-20 Continuous Composites Inc. Additive manufacturing system
US11760030B2 (en) 2020-06-23 2023-09-19 Continuous Composites Inc. Systems and methods for controlling additive manufacturing
US11813793B2 (en) 2020-09-11 2023-11-14 Continuous Composites Inc. Print head for additive manufacturing system
USD995629S1 (en) 2021-01-29 2023-08-15 Wobble Works, Inc. Drawing tool
US11926099B2 (en) 2021-04-27 2024-03-12 Continuous Composites Inc. Additive manufacturing system
DE202022101078U1 (en) * 2022-02-25 2023-06-13 Grunewald Gmbh & Co. Kg System for the generative production of components

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1166005A (en) * 1967-02-28 1969-10-01 Nippon Enlarging A Method for Enlarged Multicolor Printing and a Device therefor
US5784279A (en) * 1995-09-29 1998-07-21 Bpm Technology, Inc. Apparatus for making three-dimensional articles including moving build material reservoir and associated method
US6405095B1 (en) * 1999-05-25 2002-06-11 Nanotek Instruments, Inc. Rapid prototyping and tooling system
US6280784B1 (en) * 2000-02-10 2001-08-28 Nanotek Instruments, Inc Method for rapidly making a 3-D food object
AU2003260938A1 (en) * 2002-09-12 2004-04-30 Objet Geometries Ltd. Device, system and method for calibration in three-dimensional model printing
ATE393009T1 (en) * 2002-11-12 2008-05-15 Objet Geometries Ltd METHOD AND SYSTEM FOR PRINTING A THREE-DIMENSIONAL OBJECT
US7077334B2 (en) * 2003-04-10 2006-07-18 Massachusetts Institute Of Technology Positive pressure drop-on-demand printing
JP2005187786A (en) * 2003-12-26 2005-07-14 Fuji Photo Film Co Ltd Azo dye, coloring composition for use in image formation, ink, inkjet recording method, thermal recording material, color toner, and color filter
US7939003B2 (en) 2004-08-11 2011-05-10 Cornell Research Foundation, Inc. Modular fabrication systems and methods
JP4725577B2 (en) * 2007-12-28 2011-07-13 カシオ計算機株式会社 Manufacturing method of display device
US10259161B2 (en) 2011-01-24 2019-04-16 Cornell University Deposition of materials for edible solid freeform fabrication

Cited By (56)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150323920A1 (en) * 2013-01-18 2015-11-12 Trumpf Werkzeugmaschinen Gmbh + Co. Kg Machine Maintenance Using a Machine Controller and a Service Computer
US10671037B2 (en) * 2013-01-18 2020-06-02 Trumpf Werkzeugmaschinen Gmbh + Co. Kg Machine maintenance using a machine controller and a service computer
US10682844B2 (en) 2013-03-22 2020-06-16 Markforged, Inc. Embedding 3D printed fiber reinforcement in molded articles
US11237542B2 (en) 2013-03-22 2022-02-01 Markforged, Inc. Composite filament 3D printing using complementary reinforcement formations
US11981069B2 (en) 2013-03-22 2024-05-14 Markforged, Inc. Three dimensional printing of composite reinforced structures
US11787104B2 (en) 2013-03-22 2023-10-17 Markforged, Inc. Methods for fiber reinforced additive manufacturing
US11759990B2 (en) 2013-03-22 2023-09-19 Markforged, Inc. Three dimensional printing
US10696039B2 (en) 2013-03-22 2020-06-30 Markforged, Inc. Multilayer fiber reinforcement design for 3D printing
US11577462B2 (en) 2013-03-22 2023-02-14 Markforged, Inc. Scanning print bed and part height in 3D printing
US11504892B2 (en) 2013-03-22 2022-11-22 Markforged, Inc. Impregnation system for composite filament fabrication in three dimensional printing
US11148409B2 (en) 2013-03-22 2021-10-19 Markforged, Inc. Three dimensional printing of composite reinforced structures
US11065861B2 (en) 2013-03-22 2021-07-20 Markforged, Inc. Methods for composite filament threading in three dimensional printing
US11014305B2 (en) 2013-03-22 2021-05-25 Markforged, Inc. Mid-part in-process inspection for 3D printing
US9815268B2 (en) 2013-03-22 2017-11-14 Markforged, Inc. Multiaxis fiber reinforcement for 3D printing
US9956725B2 (en) 2013-03-22 2018-05-01 Markforged, Inc. Three dimensional printer for fiber reinforced composite filament fabrication
US10016942B2 (en) 2013-03-22 2018-07-10 Markforged, Inc. Three dimensional printing
US10040252B2 (en) 2013-03-22 2018-08-07 Markforged, Inc. Methods for fiber reinforced additive manufacturing
US10076875B2 (en) 2013-03-22 2018-09-18 Markforged, Inc. Methods for composite filament fabrication in three dimensional printing
US10076876B2 (en) 2013-03-22 2018-09-18 Markforged, Inc. Three dimensional printing
US10099427B2 (en) 2013-03-22 2018-10-16 Markforged, Inc. Three dimensional printer with composite filament fabrication
US10953610B2 (en) 2013-03-22 2021-03-23 Markforged, Inc. Three dimensional printer with composite filament fabrication
US10953609B1 (en) 2013-03-22 2021-03-23 Markforged, Inc. Scanning print bed and part height in 3D printing
US10259160B2 (en) 2013-03-22 2019-04-16 Markforged, Inc. Wear resistance in 3D printing of composites
US10434702B2 (en) 2013-03-22 2019-10-08 Markforged, Inc. Additively manufactured part including a compacted fiber reinforced composite filament
US10821662B2 (en) 2013-03-22 2020-11-03 Markforged, Inc. Methods for fiber reinforced additive manufacturing
US10603841B2 (en) 2013-03-22 2020-03-31 Markforged, Inc. Multilayer fiber reinforcement design for 3D printing
US10611082B2 (en) 2013-03-22 2020-04-07 Markforged, Inc. Apparatus for fiber reinforced additive manufacturing
US10717228B2 (en) 2013-03-22 2020-07-21 Markforged, Inc. Three dimensional printing
US11420382B2 (en) 2013-03-22 2022-08-23 Markforged, Inc. Apparatus for fiber reinforced additive manufacturing
CN105705319A (en) * 2013-09-19 2016-06-22 马克弗巨德有限公司 Methods for fiber reinforced additive manufacturing
CN103522547A (en) * 2013-09-26 2014-01-22 上海大学 Numerically-controlled machine tool power-driven 3D (three dimensional) printing head component and method for manufacturing three-dimensional support
US10564172B2 (en) 2014-05-08 2020-02-18 The Cleveland Clinic Foundation Systems and methods for detection, analysis, isolation and/or harvesting of biological objects
WO2015172025A1 (en) * 2014-05-08 2015-11-12 The Cleveland Clinic Foundation Systems and methods for detection, analysis, isolation and/or harvesting of biological objects
US11579160B2 (en) 2014-05-08 2023-02-14 The Cleveland Clinic Foundation Systems and methods for detection, analysis, isolation and/or harvesting of biological objects
US20160011083A1 (en) * 2014-07-11 2016-01-14 Intellicyt Contact Sensing Probe and Methods of Use for Microplate Liquid Sampling
US9797917B2 (en) * 2014-07-11 2017-10-24 Intellicyt Contact sensing probe and methods of use for microplate liquid sampling
CN107000318A (en) * 2014-12-01 2017-08-01 沙特基础工业全球技术有限公司 The nozzle instrument for extruding increasing material manufacturing for material changes
CN104908321A (en) * 2015-05-22 2015-09-16 杭州汉卓机电科技有限公司 Control method of pneumatic double-acting biological 3D printing nozzle with internal and external double spiral cover structures
CN106292416A (en) * 2015-06-24 2017-01-04 英国电讯有限公司 The gate printed
WO2017001676A1 (en) * 2015-07-02 2017-01-05 Hamilton Bonaduz Ag Pipetting apparatus with image processing
DE102015212457A1 (en) * 2015-07-02 2017-01-05 Hamilton Bonaduz Ag Pipetting device with image processing
CN105216312A (en) * 2015-10-12 2016-01-06 周末 A kind of 3D printing device checkout gear
US20190054624A1 (en) * 2016-03-17 2019-02-21 Flow Robotics A/S Robot and a method of controlling a robot
US10702990B2 (en) 2016-03-17 2020-07-07 Flow Robotics A/S Robot and a method of controlling a robot
WO2017157402A1 (en) * 2016-03-17 2017-09-21 Itu Business Development A/S A robot and a method of controlling a robot
US11759997B2 (en) * 2016-05-12 2023-09-19 Hewlett-Packard Development Company, L.P. Build material splash control
US20190054682A1 (en) * 2016-05-12 2019-02-21 Hewlett-Packard Development Company, L.P. Build material splash control
US11135579B2 (en) 2016-07-13 2021-10-05 Hewlett-Packard Development Company, L.P. Apparatus with encoded media to indicate dispensing locations for pipette dispenser
US11263433B2 (en) 2016-10-28 2022-03-01 Beckman Coulter, Inc. Substance preparation evaluation system
US20210181222A1 (en) * 2018-04-23 2021-06-17 Shimadzu Corporation Autosampler
US20210299870A1 (en) * 2018-08-24 2021-09-30 The University Of Tokyo Robot assistance device and robot assistance system
JP2022518066A (en) * 2019-01-28 2022-03-11 フォーミュラトリックス・インコーポレイテッド Liquid processing equipment and pipette heads and methods for sucking and / or distributing liquids
US20200408791A1 (en) * 2019-06-28 2020-12-31 Sartorius Biohit Liquid Handling Oy Method of information transmission, a liquid handling device, and a system
US11585822B2 (en) * 2019-06-28 2023-02-21 Sartorius Biohit Liquid Handling Oy Method of information transmission, a liquid handling device, and a system
WO2022013141A1 (en) * 2020-07-16 2022-01-20 Signify Holding B.V. Single camera-based process monitoring in fdm 3d printer
CN116728778A (en) * 2023-07-17 2023-09-12 北京科技大学 Basalt fiber composite material 3D printing process equipment

Also Published As

Publication number Publication date
US9884318B2 (en) 2018-02-06
US20130209600A1 (en) 2013-08-15

Similar Documents

Publication Publication Date Title
US20130205920A1 (en) Automated visual pipetting
US20200406251A1 (en) Devices and methods for programmable manipulation of pipettes
CN208367018U (en) The electromechanical assemblies of Laboratory Instruments, automated system and mobile microbiological test plate that analyte is tested
JP6290903B2 (en) Experimental equipment and method for automatic processing of experimental samples
JP5346807B2 (en) A device that automatically adjusts the bacterial inoculum level of a sample
KR20110133589A (en) Device for automatically cultivating cells in parallel
CN111093830A (en) Object picking device with imaging-based pipette tip positioning
CN111936863B (en) Automatic analysis device
EP3888857B1 (en) Cell production apparatus, cell production method, program for cell production apparatus, and cell production system
JP7013804B2 (en) Container and calibration standard plate
WO2018100917A1 (en) Information processing device, observation system, information processing method, and program
US20190257849A1 (en) Virtual pipetting
AU2019226329A1 (en) Measuring apparatus for a laboratory appliance for measuring an article, article for said measuring apparatus and measuring method
CN103954784A (en) Vaginitis automated inspection workstation
US10591501B2 (en) Automatic structure determination
WO2019159524A1 (en) Imaging system and biological subject transfer device
CN220819750U (en) Blood urine manure has component detection device and multi-parameter to have component detection chip
US12019084B2 (en) Laboratory automation device control program generation with object detection
Grexa et al. SpheroidPicker: An Automated 3D cell culture manipulator robot using deep learning
Dove Cell biology shapes up
CN117871355A (en) Blood urine manure and multiparameter formed component detection device, chip and method
Feldman et al. Proposed automation of a human-tumor-drug screen.

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- INCOMPLETE APPLICATION (PRE-EXAMINATION)