WO2024106132A1 - Solid-state imaging device and information processing system - Google Patents

Solid-state imaging device and information processing system Download PDF

Info

Publication number
WO2024106132A1
WO2024106132A1 PCT/JP2023/037961 JP2023037961W WO2024106132A1 WO 2024106132 A1 WO2024106132 A1 WO 2024106132A1 JP 2023037961 W JP2023037961 W JP 2023037961W WO 2024106132 A1 WO2024106132 A1 WO 2024106132A1
Authority
WO
WIPO (PCT)
Prior art keywords
event data
event
unit
octave
information processing
Prior art date
Application number
PCT/JP2023/037961
Other languages
French (fr)
Japanese (ja)
Inventor
元就 本田
Original Assignee
ソニーセミコンダクタソリューションズ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーセミコンダクタソリューションズ株式会社 filed Critical ソニーセミコンダクタソリューションズ株式会社
Publication of WO2024106132A1 publication Critical patent/WO2024106132A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/47Image sensors with pixel address output; Event-driven image sensors; Selection of pixels to be read out based on image data

Definitions

  • This disclosure relates to a solid-state imaging device and an information processing system.
  • an event sensor is a sensor for detecting changes in the subject.
  • EVS Event-based Vision Sensor
  • an event sensor can set a higher frame rate than an image sensor.
  • an event sensor is realized by a solid-state imaging device such as a CCD (Charge Coupled Device) sensor or a CMOS (Complementary Metal Oxide Semiconductor) sensor.
  • EVSs output event data with a certain spatial resolution.
  • an information processing system receives this event data and performs information processing such as object recognition, it performs processes such as scaling, filtering, and cropping on the event data to extract useful information from it.
  • one method that has been considered is to spatially downsample event data by dropping the least significant bits of the event address information. This reduces the spatial resolution of the event data, thereby preventing increases in data mining costs and reducing delays in information processing in information processing systems.
  • the reduction in spatial resolution may result in a decrease in the accuracy of target tasks such as object recognition.
  • the present disclosure provides a solid-state imaging device and information processing system that can achieve high resolution while suppressing delays in information processing.
  • the solid-state imaging device includes a plurality of pixels that detect an event and output event data indicating the detection result of the event, an event generation unit that converts the event data output from the plurality of pixels into event data of a first octave and generates event data of a second to i-th octave (i is an integer of 2 or more) from the event data of the first octave, and an event output unit that outputs at least a portion of the event data of the first to i-th octave.
  • the solid-state imaging device of the first aspect may further include an octave information adding unit that adds octave information, which is identification information for the first to i-th octaves, to the event data of the first to i-th octaves, respectively, and the event output unit may output the event data to which the octave information has been added. This makes it possible, for example, to identify which octave a certain event data belongs to.
  • the event output unit may output the event data in an image representation for each octave. This makes it possible to display event data for various octaves as images, for example.
  • the event generating unit may generate event data for the j+1th octave from event data for the jth octave (j is an integer satisfying 1 ⁇ j ⁇ i ⁇ 1). This makes it possible to sequentially generate event data for various octaves from event data for the first octave, for example, by generating event data for the second octave from event data for the first octave, and generating event data for the third octave from event data for the second octave.
  • the event generation unit may generate one line of event data for the j+1 octave from m lines (m is an integer equal to or greater than 2) of event data for the j octave. This makes it possible to easily generate event data for the j+1 octave from event data for the j octave, for example.
  • the event generation unit may generate one row and one column of event data for the j+1 octave from m rows and n columns (n is an integer of 2 or more) of event data for the j octave. This makes it possible to more easily generate event data for the j+1 octave from event data for the j octave, for example.
  • the event generating unit may ignite an event with the event data of the j+1 octave of the 1 row and 1 column when the event data of the j octave of the m rows and n columns includes k event firings (k is an integer satisfying 1 ⁇ k ⁇ m ⁇ n). This makes it possible to easily replace, for example, m ⁇ n areas of the j th octave with one area of the j+1 th octave.
  • m may be 2
  • n may be 2
  • k may be 1, 2, 3, and 4. This makes it possible to distinguish, for example, between a case where no events are fired in the m ⁇ n regions of the jth octave and a case where at least one event is fired in the m ⁇ n regions of the jth octave.
  • m may be 2
  • n may be 2
  • k may be 2, 3, or 4. This makes it possible to suppress the effects of noise events, for example.
  • m may be 2
  • n may be 2
  • k may be 3 and 4. This makes it possible to suppress the effects of noise events and flicker events, for example.
  • the solid-state imaging device of the first aspect may further include a frame memory that stores the event data output from the plurality of pixels, and the event generating unit may treat the event data output from the frame memory as event data of the first octave. This makes it possible to employ, for example, an arbiter-type event sensor as the solid-state imaging device.
  • the information processing system is an information processing system including a solid-state imaging device and an information processing unit, in which the solid-state imaging device includes a plurality of pixels that detect an event and output event data indicating the detection result of the event, an event generation unit that treats the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer of 2 or more) from the event data of the first octave, and an event output unit that outputs at least a portion of the event data of the first to i-th octave, and the information processing unit displays the event data output from the event output unit on a display screen.
  • the solid-state imaging device includes a plurality of pixels that detect an event and output event data indicating the detection result of the event
  • an event generation unit that treats the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to
  • the information processing unit may include an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, and the information processing unit may display the event data extracted by the extraction unit on the display screen. This makes it possible, for example, to use the event data of various octaves output from a solid-state imaging device for display on an octave basis.
  • the information processing unit may include an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, and the information processing unit may perform image recognition using the event data extracted by the extraction unit.
  • the image recognition may be user gesture recognition. This makes it possible, for example, to use event data of various octaves output from a solid-state imaging device for gesture recognition for each octave.
  • the information processing unit may include a selection unit that selects the event data for a number of octaves specified by a user from the event data output from the event output unit, and the information processing unit may record the event data selected by the selection unit on a recording medium. This makes it possible to record, for example, event data for various octaves output from a solid-state imaging device for each octave.
  • the information processing system may be an electronic device including the solid-state imaging device and the information processing unit. This makes it possible, for example, to perform the output of event data and the subsequent information processing in the same electronic device.
  • the electronic device may further include a display unit having the display screen. This makes it possible, for example, to output the event data and display the event data on the same electronic device.
  • the information processing system may include an electronic device including the information processing unit, and an imaging device that is provided outside the electronic device and includes the solid-state imaging device. This makes it possible, for example, to perform information processing using event data in an electronic device outside the imaging device.
  • the information processing system may further include a display device provided outside the electronic device and having the display screen. This makes it possible, for example, to display event data on an electronic device external to the imaging device.
  • FIG. 1 is a block diagram showing a configuration of a vehicle 1 according to a first embodiment.
  • FIG. 2 is a plan view showing a sensing area of the vehicle 1 according to the first embodiment.
  • 1 is a block diagram showing a configuration of a solid-state imaging device 100 according to a first embodiment.
  • 2 is a diagram for explaining a pixel array 101 according to the first embodiment.
  • FIG. 4A to 4C are diagrams for explaining the operation of the solid-state imaging device 100 according to the first embodiment.
  • 4 is a diagram for explaining the operation of an event generating unit 103 in the first embodiment.
  • FIG. FIG. 11 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
  • FIG. 11 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
  • FIG. 4A to 4C are diagrams for explaining the operation of an event output unit 105 according to the first embodiment.
  • FIG. 1 illustrates an example of an electronic device 200 according to a first embodiment.
  • FIG. 2 is a diagram illustrating another example of the electronic device 200 according to the first embodiment.
  • FIG. 12 is a diagram for explaining details of the electronic device 200 shown in FIG. 11 .
  • FIG. 11 is a block diagram showing a configuration of a solid-state imaging device 100 according to a second embodiment.
  • FIG. 11 is a diagram for explaining the operation of an event generating unit 103 in the second embodiment.
  • FIG. 11 is a perspective view illustrating a schematic configuration of a solid-state imaging device 100 according to a third embodiment.
  • FIG. 13 is a plan view illustrating a schematic configuration of a photosensor chip 120 according to a third embodiment.
  • FIG. 13 is a plan view illustrating a schematic configuration of a detection chip 130 according to a third embodiment.
  • FIG. 13 is a circuit diagram showing a configuration of each address event detection circuit 131a according to a third embodiment.
  • FIG. 11 is a circuit diagram showing a configuration of a current-voltage conversion circuit 310 according to a third embodiment.
  • FIG. 11 is a circuit diagram showing configurations of a subtractor 330 and a quantizer 340 according to a third embodiment.
  • FIG. 13 is a circuit diagram showing the configurations of the light receiving chip 120 and the detection chip 130 according to a modified example of the third embodiment.
  • Vehicle 1 of the First Embodiment Fig. 1 is a block diagram showing the configuration of a vehicle 1 according to the first embodiment.
  • Fig. 1 shows an example of the configuration of a vehicle control system 11, which is an example of a mobility device control system.
  • the vehicle control system 11 is installed in the vehicle 1 and performs processing related to driving assistance and autonomous driving of the vehicle 1.
  • the vehicle control system 11 includes a vehicle control ECU (Electronic Control Unit) 21, a communication unit 22, a map information storage unit 23, a location information acquisition unit 24, an external recognition sensor 25, an in-vehicle sensor 26, a vehicle sensor 27, a memory unit 31, a driving assistance/automated driving control unit 32, a DMS (Driver Monitoring System) 33, an HMI (Human Machine Interface) 34, and a vehicle control unit 35.
  • vehicle control ECU Electronic Control Unit
  • communication unit 22 includes a communication unit 22, a map information storage unit 23, a location information acquisition unit 24, an external recognition sensor 25, an in-vehicle sensor 26, a vehicle sensor 27, a memory unit 31, a driving assistance/automated driving control unit 32, a DMS (Driver Monitoring System) 33, an HMI (Human Machine Interface) 34, and a vehicle control unit 35.
  • the vehicle control ECU 21, communication unit 22, map information storage unit 23, position information acquisition unit 24, external recognition sensor 25, in-vehicle sensor 26, vehicle sensor 27, memory unit 31, driving assistance/automatic driving control unit 32, driver monitoring system (DMS) 33, human machine interface (HMI) 34, and vehicle control unit 35 are connected to each other so as to be able to communicate with each other via a communication network 41.
  • the communication network 41 is composed of an in-vehicle communication network or bus that complies with a digital two-way communication standard such as CAN (Controller Area Network), LIN (Local Interconnect Network), LAN (Local Area Network), FlexRay (registered trademark), or Ethernet (registered trademark).
  • the communication network 41 may be used differently depending on the type of data being transmitted.
  • CAN may be applied to data related to vehicle control
  • Ethernet may be applied to large-volume data.
  • each part of the vehicle control system 11 may be directly connected without going through the communication network 41, using wireless communication intended for communication over relatively short distances, such as near field communication (NFC) or Bluetooth (registered trademark).
  • NFC near field communication
  • Bluetooth registered trademark
  • the vehicle control ECU 21 is configured with various processors such as a CPU (Central Processing Unit), an MPU (Micro Processing Unit), etc.
  • the vehicle control ECU 21 controls the entire or part of the functions of the vehicle control system 11.
  • the communication unit 22 communicates with various devices inside and outside the vehicle, other vehicles, servers, base stations, etc., and transmits and receives various types of data. At this time, the communication unit 22 can communicate using a plurality of communication methods.
  • the communication unit 22 communicates with servers (hereinafter referred to as external servers) on an external network via base stations or access points using wireless communication methods such as 5G (fifth generation mobile communication system), LTE (Long Term Evolution), and DSRC (Dedicated Short Range Communications).
  • the external network with which the communication unit 22 communicates is, for example, the Internet, a cloud network, or an operator-specific network.
  • the communication method that the communication unit 22 uses with the external network is not particularly limited as long as it is a wireless communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined distance.
  • the communication unit 22 can communicate with a terminal present in the vicinity of the vehicle using P2P (Peer To Peer) technology.
  • the terminal present in the vicinity of the vehicle can be, for example, a terminal attached to a mobile object moving at a relatively slow speed, such as a pedestrian or a bicycle, a terminal installed at a fixed position in a store, or an MTC (Machine Type Communication) terminal.
  • the communication unit 22 can also perform V2X communication.
  • V2X communication refers to communication between the vehicle and others, such as vehicle-to-vehicle communication with other vehicles, vehicle-to-infrastructure communication with roadside devices, vehicle-to-home communication with a home, and vehicle-to-pedestrian communication with a terminal carried by a pedestrian, etc.
  • the communication unit 22 can, for example, receive from the outside a program for updating the software that controls the operation of the vehicle control system 11 (Over the Air).
  • the communication unit 22 can further receive map information, traffic information, information about the surroundings of the vehicle 1, etc. from the outside.
  • the communication unit 22 can also transmit information about the vehicle 1 and information about the surroundings of the vehicle 1 to the outside.
  • Information about the vehicle 1 that the communication unit 22 transmits to the outside includes, for example, data indicating the state of the vehicle 1, the recognition results by the recognition unit 73, etc.
  • the communication unit 22 performs communication corresponding to a vehicle emergency notification system such as e-Call.
  • the communication unit 22 receives electromagnetic waves transmitted by a road traffic information and communication system (VICS (Vehicle Information and Communication System) (registered trademark)) such as a radio beacon, optical beacon, or FM multiplex broadcasting.
  • VICS Vehicle Information and Communication System
  • the communication unit 22 can communicate with each device in the vehicle using, for example, wireless communication.
  • the communication unit 22 can perform wireless communication with each device in the vehicle using a communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined speed via wireless communication, such as wireless LAN, Bluetooth, NFC, or WUSB (Wireless USB).
  • the communication unit 22 can also communicate with each device in the vehicle using wired communication.
  • the communication unit 22 can communicate with each device in the vehicle using wired communication via a cable connected to a connection terminal (not shown).
  • the communication unit 22 can communicate with each device in the vehicle using a communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined speed via wired communication, such as USB (Universal Serial Bus), HDMI (High-Definition Multimedia Interface) (registered trademark), or MHL (Mobile High-definition Link).
  • a communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined speed via wired communication, such as USB (Universal Serial Bus), HDMI (High-Definition Multimedia Interface) (registered trademark), or MHL (Mobile High-definition Link).
  • devices in the vehicle refers to devices that are not connected to the communication network 41 in the vehicle.
  • Examples of devices in the vehicle include mobile devices and wearable devices carried by passengers such as the driver, and information devices that are brought into the vehicle and temporarily installed.
  • the map information storage unit 23 stores one or both of a map acquired from an external source and a map created by the vehicle 1.
  • the map information storage unit 23 stores a three-dimensional high-precision map, a global map that has lower precision than a high-precision map and covers a wide area, and the like.
  • High-precision maps include, for example, dynamic maps, point cloud maps, and vector maps.
  • a dynamic map is, for example, a map consisting of four layers of dynamic information, semi-dynamic information, semi-static information, and static information, and is provided to the vehicle 1 from an external server or the like.
  • a point cloud map is a map composed of a point cloud (point group data).
  • a vector map is, for example, a map that associates traffic information such as the positions of lanes and traffic lights with a point cloud map, and is adapted for ADAS (Advanced Driver Assistance System) and AD (Autonomous Driving).
  • the point cloud map and vector map may be provided, for example, from an external server, or may be created by the vehicle 1 based on sensing results from the camera 51, radar 52, LiDAR 53, etc. as a map for matching with a local map described below, and stored in the map information storage unit 23.
  • map data of, for example, an area of several hundred meters square regarding the planned route along which the vehicle 1 will travel is acquired from the external server, etc., in order to reduce communication capacity.
  • the position information acquisition unit 24 receives GNSS signals from Global Navigation Satellite System (GNSS) satellites and acquires position information of the vehicle 1.
  • GNSS Global Navigation Satellite System
  • the acquired position information is supplied to the driving assistance/automated driving control unit 32.
  • the position information acquisition unit 24 is not limited to a method using GNSS signals, and may acquire position information using a beacon, for example.
  • the external recognition sensor 25 includes various sensors used to recognize the situation outside the vehicle 1, and supplies sensor data from each sensor to each unit of the vehicle control system 11.
  • the type and number of sensors included in the external recognition sensor 25 are arbitrary.
  • the external recognition sensor 25 includes a camera 51, a radar 52, a LiDAR (Light Detection and Ranging, Laser Imaging Detection and Ranging) 53, and an ultrasonic sensor 54.
  • the external recognition sensor 25 may be configured to include one or more types of sensors among the camera 51, the radar 52, the LiDAR 53, and the ultrasonic sensor 54.
  • the number of cameras 51, radars 52, LiDAR 53, and ultrasonic sensors 54 is not particularly limited as long as it is a number that can be realistically installed on the vehicle 1.
  • the types of sensors included in the external recognition sensor 25 are not limited to this example, and the external recognition sensor 25 may include other types of sensors. Examples of the sensing areas of each sensor included in the external recognition sensor 25 will be described later.
  • the imaging method of camera 51 is not particularly limited.
  • cameras of various imaging methods such as a ToF (Time Of Flight) camera, a stereo camera, a monocular camera, and an infrared camera, which are imaging methods capable of distance measurement, can be applied to camera 51 as necessary.
  • ToF Time Of Flight
  • stereo camera stereo camera
  • monocular camera stereo camera
  • infrared camera infrared camera
  • the present invention is not limited to this, and camera 51 may simply be used to obtain a photographed image, without being related to distance measurement.
  • the external recognition sensor 25 can be equipped with an environmental sensor for detecting the environment relative to the vehicle 1.
  • the environmental sensor is a sensor for detecting the environment such as the weather, climate, brightness, etc., and can include various sensors such as a raindrop sensor, a fog sensor, a sunlight sensor, a snow sensor, an illuminance sensor, etc.
  • the external recognition sensor 25 includes a microphone that is used to detect sounds around the vehicle 1 and the location of sound sources.
  • the in-vehicle sensor 26 includes various sensors for detecting information inside the vehicle, and supplies sensor data from each sensor to each unit of the vehicle control system 11.
  • the types and number of the various sensors included in the in-vehicle sensor 26 are not particularly limited as long as they are of the types and number that can be realistically installed in the vehicle 1.
  • the in-vehicle sensor 26 may be equipped with one or more types of sensors including a camera, radar, a seating sensor, a steering wheel sensor, a microphone, and a biometric sensor.
  • the camera equipped in the in-vehicle sensor 26 may be a camera using various imaging methods capable of measuring distances, such as a ToF camera, a stereo camera, a monocular camera, or an infrared camera. Without being limited to this, the camera equipped in the in-vehicle sensor 26 may be a camera simply for acquiring captured images, regardless of distance measurement.
  • the biometric sensor equipped in the in-vehicle sensor 26 is provided, for example, on a seat, steering wheel, etc., and detects various types of biometric information of passengers such as the driver.
  • the vehicle sensor 27 includes various sensors for detecting the state of the vehicle 1, and supplies sensor data from each sensor to each unit of the vehicle control system 11.
  • the types and number of the various sensors included in the vehicle sensor 27 are not particularly limited as long as they are types and numbers that can be realistically installed in the vehicle 1.
  • the vehicle sensor 27 includes a speed sensor, an acceleration sensor, an angular velocity sensor (gyro sensor), and an inertial measurement unit (IMU) that integrates these.
  • the vehicle sensor 27 includes a steering angle sensor that detects the steering angle of the steering wheel, a yaw rate sensor, an accelerator sensor that detects the amount of accelerator pedal operation, and a brake sensor that detects the amount of brake pedal operation.
  • the vehicle sensor 27 includes a rotation sensor that detects the number of rotations of the engine or motor, an air pressure sensor that detects the air pressure of the tires, a slip ratio sensor that detects the slip ratio of the tires, and a wheel speed sensor that detects the rotation speed of the wheels.
  • the vehicle sensor 27 includes a battery sensor that detects the remaining charge and temperature of the battery, and an impact sensor that detects external impacts.
  • the storage unit 31 includes at least one of a non-volatile storage medium and a volatile storage medium, and stores data and programs.
  • the storage unit 31 is used, for example, as an electrically erasable programmable read only memory (EEPROM) and a random access memory (RAM), and the storage medium may be a magnetic storage device such as a hard disk drive (HDD), a semiconductor storage device, an optical storage device, or a magneto-optical storage device.
  • the storage unit 31 stores various programs and data used by each part of the vehicle control system 11.
  • the storage unit 31 includes an event data recorder (EDR) and a data storage system for automated driving (DSSAD), and stores information on the vehicle 1 before and after an event such as an accident, and information acquired by the in-vehicle sensor 26.
  • EDR event data recorder
  • DSSAD data storage system for automated driving
  • the driving assistance/automatic driving control unit 32 controls driving assistance and automatic driving of the vehicle 1.
  • the driving assistance/automatic driving control unit 32 includes an analysis unit 61, an action planning unit 62, and an operation control unit 63.
  • the analysis unit 61 performs analysis processing of the vehicle 1 and the surrounding conditions.
  • the analysis unit 61 includes a self-position estimation unit 71, a sensor fusion unit 72, and a recognition unit 73.
  • the self-position estimation unit 71 estimates the self-position of the vehicle 1 based on the sensor data from the external recognition sensor 25 and the high-precision map stored in the map information storage unit 23. For example, the self-position estimation unit 71 generates a local map based on the sensor data from the external recognition sensor 25, and estimates the self-position of the vehicle 1 by matching the local map with the high-precision map.
  • the position of the vehicle 1 is based on, for example, the center of the rear wheel pair axle.
  • the local map is, for example, a three-dimensional high-precision map or an occupancy grid map created using technology such as SLAM (Simultaneous Localization and Mapping).
  • the three-dimensional high-precision map is, for example, the point cloud map described above.
  • the occupancy grid map is a map in which the three-dimensional or two-dimensional space around the vehicle 1 is divided into grids of a predetermined size, and the occupancy state of objects is shown on a grid-by-grid basis.
  • the occupancy state of objects is indicated, for example, by the presence or absence of an object and the probability of its existence.
  • the local map is also used, for example, in detection processing and recognition processing of the situation outside the vehicle 1 by the recognition unit 73.
  • the self-position estimation unit 71 may estimate the self-position of the vehicle 1 based on the position information acquired by the position information acquisition unit 24 and the sensor data from the vehicle sensor 27.
  • the sensor fusion unit 72 performs sensor fusion processing to combine multiple different types of sensor data (e.g., image data supplied from the camera 51 and sensor data supplied from the radar 52) to obtain new information.
  • Methods for combining different types of sensor data include integration, fusion, and association.
  • the recognition unit 73 executes a detection process to detect the situation outside the vehicle 1, and a recognition process to recognize the situation outside the vehicle 1.
  • the recognition unit 73 performs detection and recognition processing of the situation outside the vehicle 1 based on information from the external recognition sensor 25, information from the self-position estimation unit 71, information from the sensor fusion unit 72, etc.
  • the recognition unit 73 performs detection processing and recognition processing of objects around the vehicle 1.
  • Object detection processing is, for example, processing to detect the presence or absence, size, shape, position, movement, etc. of an object.
  • Object recognition processing is, for example, processing to recognize attributes such as the type of object, and to identify a specific object.
  • detection processing and recognition processing are not necessarily clearly separated, and there may be overlap.
  • the recognition unit 73 detects objects around the vehicle 1 by performing clustering to classify a point cloud based on sensor data from the radar 52, the LiDAR 53, or the like into clusters of points. This allows the presence or absence, size, shape, and position of objects around the vehicle 1 to be detected.
  • the recognition unit 73 detects the movement of objects around the vehicle 1 by performing tracking to follow the movement of clusters of point clouds classified by clustering. This allows the speed and direction of travel (movement vector) of objects around the vehicle 1 to be detected.
  • the recognition unit 73 detects or recognizes vehicles, people, bicycles, obstacles, structures, roads, traffic lights, traffic signs, road markings, etc. based on image data supplied from the camera 51.
  • the recognition unit 73 may also recognize the types of objects around the vehicle 1 by performing recognition processing such as semantic segmentation.
  • the recognition unit 73 can perform recognition processing of traffic rules around the vehicle 1 based on the map stored in the map information storage unit 23, the result of self-location estimation by the self-location estimation unit 71, and the result of recognition of objects around the vehicle 1 by the recognition unit 73. Through this processing, the recognition unit 73 can recognize the positions and states of traffic lights, the contents of traffic signs and road markings, the contents of traffic regulations, and lanes on which travel is possible, etc.
  • the recognition unit 73 can perform recognition processing of the environment around the vehicle 1.
  • the surrounding environment that the recognition unit 73 recognizes may include weather, temperature, humidity, brightness, and road surface conditions.
  • the behavior planning unit 62 creates a behavior plan for the vehicle 1. For example, the behavior planning unit 62 creates the behavior plan by performing route planning and route following processing.
  • Global path planning is a process that plans a rough route from the start to the goal. This route planning is called trajectory planning, and also includes a process of local path planning that takes into account the motion characteristics of vehicle 1 on the planned route and generates a trajectory that allows safe and smooth progress in the vicinity of vehicle 1.
  • Path following is a process of planning operations for traveling safely and accurately along a route planned by a route plan within a planned time.
  • the action planning unit 62 can, for example, calculate the target speed and target angular velocity of the vehicle 1 based on the results of this path following process.
  • the operation control unit 63 controls the operation of the vehicle 1 to realize the action plan created by the action planning unit 62.
  • the operation control unit 63 controls the steering control unit 81, the brake control unit 82, and the drive control unit 83 included in the vehicle control unit 35 described below, and performs acceleration/deceleration control and directional control so that the vehicle 1 proceeds along the trajectory calculated by the trajectory plan.
  • the operation control unit 63 performs cooperative control aimed at realizing ADAS functions such as collision avoidance or impact mitigation, following driving, maintaining vehicle speed, collision warning for the vehicle itself, and lane departure warning for the vehicle itself.
  • the operation control unit 63 performs cooperative control aimed at automatic driving, which drives autonomously without the driver's operation.
  • the DMS 33 performs authentication processing of the driver and recognition processing of the driver's state based on the sensor data from the in-vehicle sensor 26 and input data input to the HMI 34 (described later), etc.
  • Examples of the driver's state to be recognized include physical condition, alertness level, concentration level, fatigue level, line of sight direction, level of intoxication, driving operation, posture, etc.
  • the DMS 33 may also perform authentication processing for passengers other than the driver and recognition processing for the status of the passengers.
  • the DMS 33 may also perform recognition processing for the situation inside the vehicle based on sensor data from the in-vehicle sensor 26. Examples of the situation inside the vehicle that may be recognized include temperature, humidity, brightness, odor, etc.
  • HMI34 The HMI 34 inputs various data and instructions, and presents various data to the driver, etc.
  • the HMI 34 is equipped with an input device that allows a person to input data.
  • the HMI 34 generates input signals based on data and instructions input by the input device, and supplies the signals to each part of the vehicle control system 11.
  • the HMI 34 is equipped with input devices such as a touch panel, buttons, switches, and levers. Without being limited to these, the HMI 34 may further be equipped with an input device that allows information to be input by a method other than manual operation, such as voice or gestures.
  • the HMI 34 may use, as an input device, an externally connected device such as a remote control device that uses infrared rays or radio waves, or a mobile device or wearable device that supports the operation of the vehicle control system 11.
  • the HMI 34 generates visual information, auditory information, and tactile information for the occupant or the outside of the vehicle.
  • the HMI 34 also performs output control to control the output, output content, output timing, output method, etc. of each piece of generated information.
  • the HMI 34 generates and outputs, as visual information, information indicated by images or light, such as an operation screen, a status display of the vehicle 1, a warning display, and a monitor image showing the situation around the vehicle 1.
  • the HMI 34 also generates and outputs, as auditory information, information indicated by sounds, such as voice guidance, warning sounds, and warning messages.
  • the HMI 34 also generates and outputs, as tactile information, information that is imparted to the occupant's sense of touch by, for example, force, vibration, movement, etc.
  • the output device from which the HMI 34 outputs visual information may be, for example, a display device that presents visual information by displaying an image itself, or a projector device that presents visual information by projecting an image.
  • the display device may be a device that displays visual information within the field of vision of the passenger, such as a head-up display, a transmissive display, or a wearable device with an AR (Augmented Reality) function, in addition to a display device having a normal display.
  • the HMI 34 may also use display devices such as a navigation device, an instrument panel, a CMS (Camera Monitoring System), an electronic mirror, or lamps provided in the vehicle 1 as output devices that output visual information.
  • the output device through which the HMI 34 outputs auditory information can be, for example, an audio speaker, headphones, or earphones.
  • Haptic elements using haptic technology can be used as output devices for the HMI 34 to output haptic information.
  • Haptic elements are provided on parts of the vehicle 1 that are in contact with passengers, such as the steering wheel and the seat.
  • the vehicle control unit 35 controls each unit of the vehicle 1.
  • the vehicle control unit 35 includes a steering control unit 81, a brake control unit 82, a drive control unit 83, a body system control unit 84, a light control unit 85, and a horn control unit 86.
  • the steering control unit 81 detects and controls the state of the steering system of the vehicle 1.
  • the steering system includes, for example, a steering mechanism including a steering wheel, an electric power steering, etc.
  • the steering control unit 81 includes, for example, a steering ECU that controls the steering system, an actuator that drives the steering system, etc.
  • the brake control unit 82 detects and controls the state of the brake system of the vehicle 1.
  • the brake system includes, for example, a brake mechanism including a brake pedal, an ABS (Antilock Brake System), a regenerative brake mechanism, etc.
  • the brake control unit 82 includes, for example, a brake ECU that controls the brake system, and an actuator that drives the brake system.
  • the drive control unit 83 detects and controls the state of the drive system of the vehicle 1.
  • the drive system includes, for example, an accelerator pedal, a drive force generating device for generating drive force such as an internal combustion engine or a drive motor, and a drive force transmission mechanism for transmitting the drive force to the wheels.
  • the drive control unit 83 includes, for example, a drive ECU for controlling the drive system, and an actuator for driving the drive system.
  • the body system control unit 84 detects and controls the state of the body system of the vehicle 1.
  • the body system includes, for example, a keyless entry system, a smart key system, a power window device, a power seat, an air conditioning system, an airbag, a seat belt, a shift lever, etc.
  • the body system control unit 84 includes, for example, a body system ECU that controls the body system, an actuator that drives the body system, etc.
  • the light control unit 85 detects and controls the state of various lights of the vehicle 1. Examples of lights to be controlled include headlights, backlights, fog lights, turn signals, brake lights, projection, and bumper displays.
  • the light control unit 85 includes a light ECU that controls the lights, an actuator that drives the lights, and the like.
  • the horn control unit 86 detects and controls the state of the car horn of the vehicle 1.
  • the horn control unit 86 includes, for example, a horn ECU that controls the car horn, an actuator that drives the car horn, etc.
  • FIG. 2 is a plan view showing the sensing area of the vehicle 1 in the first embodiment.
  • FIG. 2 shows an example of the sensing area of the camera 51, radar 52, LiDAR 53, ultrasonic sensor 54, etc. of the external recognition sensor 25 in FIG. 1.
  • FIG. 2 shows a schematic view of the vehicle 1 as seen from above, with the left end side being the front end of the vehicle 1 and the right end side being the rear end of the vehicle 1.
  • Sensing area 1-1F and sensing area 1-1B are examples of sensing areas of the ultrasonic sensors 54.
  • the sensing area 1-1F covers the periphery of the front end of the vehicle 1 with a plurality of ultrasonic sensors 54.
  • the sensing area 1-1B covers the periphery of the rear end of the vehicle 1 with a plurality of ultrasonic sensors 54.
  • sensing results in sensing area 1-1F and sensing area 1-1B are used, for example, for parking assistance for vehicle 1.
  • Sensing area 1-2F to sensing area 1-2B show examples of sensing areas of a short-range or medium-range radar 52.
  • Sensing area 1-2F covers a position farther in front of the vehicle 1 than sensing area 1-1F.
  • Sensing area 1-2B covers a position farther in the rear of the vehicle 1 than sensing area 1-1B.
  • Sensing area 1-2L covers the rear periphery of the left side of the vehicle 1.
  • Sensing area 1-2R covers the rear periphery of the right side of the vehicle 1.
  • sensing results in sensing area 1-2F are used, for example, to detect vehicles, pedestrians, etc. in front of vehicle 1.
  • the sensing results in sensing area 1-2B are used, for example, for collision prevention functions behind vehicle 1.
  • the sensing results in sensing area 1-2L and sensing area 1-2R are used, for example, to detect objects in blind spots to the sides of vehicle 1.
  • Sensing area 1-3F to sensing area 1-3B show examples of sensing areas sensed by camera 51. Sensing area 1-3F covers a position farther in front of vehicle 1 than sensing area 1-2F. Sensing area 1-3B covers a position farther in the rear of vehicle 1 than sensing area 1-2B. Sensing area 1-3L covers the periphery of the left side of vehicle 1. Sensing area 1-3R covers the periphery of the right side of vehicle 1.
  • the sensing results in sensing area 1-3F can be used, for example, for recognizing traffic lights and traffic signs, lane departure prevention support systems, and automatic headlight control systems.
  • the sensing results in sensing area 1-3B can be used, for example, for parking assistance and surround view systems.
  • the sensing results in sensing area 1-3L and sensing area 1-3R can be used, for example, for surround view systems.
  • a sensing area 1-4 shows an example of a sensing area of the LiDAR 53.
  • the sensing area 1-4 covers a position farther in front of the vehicle 1 than the sensing area 1-3F.
  • the sensing area 1-4 has a narrower range in the left-right direction than the sensing area 1-3F.
  • the sensing results in sensing areas 1-4 are used, for example, to detect objects such as nearby vehicles.
  • a sensing area 1-5 shows an example of a sensing area of a long-range radar 52.
  • the sensing area 1-5 covers a position farther ahead of the vehicle 1 than the sensing area 1-4.
  • the sensing area 1-5 has a narrower range in the left-right direction than the sensing area 1-4.
  • sensing results in sensing areas 1-5 are used, for example, for ACC (Adaptive Cruise Control), emergency braking, collision avoidance, etc.
  • ACC Adaptive Cruise Control
  • emergency braking braking
  • collision avoidance etc.
  • the sensing areas of the cameras 51, radar 52, LiDAR 53, and ultrasonic sensors 54 included in the external recognition sensor 25 may have various configurations other than those shown in FIG. 2. Specifically, the ultrasonic sensor 54 may also sense the sides of the vehicle 1, and the LiDAR 53 may sense the rear of the vehicle 1.
  • the installation positions of the sensors are not limited to the examples described above. The number of sensors may be one or more.
  • FIG. 3 is a block diagram showing the configuration of the solid-state imaging device 100 according to the first embodiment.
  • the solid-state imaging device 100 is provided in the vehicle 1 shown in FIG. 1, and is included in the external recognition sensor 25, for example.
  • the solid-state imaging device 100 is an EVS for detecting changes in a subject. Examples of subjects include a person, a vehicle, and an obstacle in front of the vehicle 1.
  • the solid-state imaging device 100 may be built into an electronic device 200 such as a smartphone, as in an example described later, or may be electrically connected to an electronic device 200 such as a game console (see FIGS. 10 and 11).
  • the solid-state imaging device 100 includes a pixel array 101, an event acquisition unit 102, an event generation unit 103, an event synthesis unit 104, and an event output unit 105.
  • the pixel array 101 includes a plurality of pixels 101a.
  • the event generation unit 103 includes a first filter unit 103a, a second filter unit 103b, and a third filter unit 103c.
  • the event synthesis unit 104 includes an octave information addition unit 104a and an output timing adjustment unit 104b.
  • the event output unit 105 includes an event data selection unit 105a and an event data formation unit 105b.
  • the pixel array 101 includes a plurality of pixels 101a arranged in a two-dimensional array (matrix).
  • matrix two-dimensional array
  • the horizontal direction on the paper corresponds to the row direction of the pixel array 101
  • the vertical direction on the paper corresponds to the column direction of the pixel array 101.
  • Each pixel 101a has the function of detecting events such as on events and off events.
  • An on event is fired when the luminance of the pixel 101a increases and the absolute value of the amount of change (increase) in luminance is greater than a threshold value.
  • An off event is fired when the luminance of the pixel 101a decreases and the absolute value of the amount of change (decrease) in luminance is greater than a threshold value. For example, an on event is fired when a subject enters the pixel 101a, and an off event is fired when a subject leaves the pixel 101a.
  • Each pixel 101a then outputs event data indicating the event detection result.
  • the event acquiring unit 102 acquires an event (event data) from each pixel 101a in the pixel array 101.
  • the event acquiring unit 102 since the solid-state imaging device 100 is a scan-type EVS, the event acquiring unit 102 sequentially scans multiple rows of the pixel array 101 and acquires event data from the pixel array 101 for each row.
  • the event data for each row of the pixel array 101 is treated as event data V1 of the first octave.
  • the event acquisition unit 102 outputs the event data V1 of the first octave to the event generation unit 103 and the event synthesis unit 104.
  • the event data V1 of the first octave acquired by the event acquisition unit 102 is sent to the event synthesis unit 104, and a copy of the event data V1 is sent to the event generation unit 103.
  • FIG. 3 shows the first octave event data V1 as a series of multiple regions P1 arranged in a row.
  • Each region P1 represents the event data for one pixel 101a in the first octave event data V1.
  • four regions P1 correspond to event data for four pixels 101a.
  • the first octave event data V1 is also called the first octave event firing/non-firing column.
  • the first octave event data V1 is character string data representing the event data for one row of pixels 101a. In this embodiment, it is possible to know whether an event has fired at a pixel 101a by acquiring the event data for that pixel 101a.
  • the event generating unit 103 regards the event data output from the event acquiring unit 102 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves (i is an integer equal to or greater than 2) from the event data V1 of the first octave.
  • the first filter unit 103a When the first filter unit 103a receives the first octave event data V1, it waits until m1 rows ( m1 is an integer equal to or greater than 2) of the first octave event data V1 are accumulated.
  • the m1 rows of the first octave event data V1 correspond to the event data of m1 rows of the pixels 101a in the pixel array 101.
  • the first filter unit 103a When the m1 rows of the first octave event data V1 are accumulated, the first filter unit 103a generates r1 rows and s1 columns of the second octave event data V2 ( r1 is an integer that satisfies r1 ⁇ m1 , s1 is an integer that satisfies s1 ⁇ n1 ) from the m1 rows and n1 columns of the first octave event data V1 ( n1 is an integer equal to or greater than 2). The first filter unit 103a then discards the m1 rows and n1 columns of the first octave event data V1. The first filter unit 103a repeats this process for the event data V1 of the first octave of all rows in order.
  • the first filter unit 103a outputs the second octave event data V2 of r 1 rows and s 1 columns to the second filter unit 103b and the event synthesis unit 104.
  • the second octave event data V2 generated by the first filter unit 103a is sent to the event synthesis unit 104, and a copy thereof is sent to the second filter unit 103b.
  • FIG. 3 shows the second octave event data V2 as a schematic representation of a plurality of regions P2 arranged in a row.
  • the pixel array 101 of this embodiment includes pixels 101a of M rows and N columns. Therefore, the process of generating the event data V2 of the second octave from the event data V1 of the first octave is repeated M ⁇ N/4 times in total.
  • FIG. 3 shows the second octave event data V2 as a schematic representation of a plurality of regions P2 arranged in a row.
  • each region P2 corresponds to event data obtained by aggregating the event data of four pixels 101a.
  • the event data V2 of the second octave is also called the event ignition presence/absence column of the second octave.
  • the second filter unit 103b When the second filter unit 103b receives the second octave event data V2, it waits until m 2 rows (m 2 is an integer equal to or greater than 2) of the second octave event data V2 are accumulated. When m 2 rows of the first octave event data V2 are accumulated, the second filter unit 103b generates r 2 rows and s 2 columns (r 2 is an integer that satisfies r 2 ⁇ m 2 , and s 2 is an integer that satisfies s 2 ⁇ n 2 ) of the third octave event data V3 from m 2 rows and n 2 columns (n 2 is an integer equal to or greater than 2 ) of the second octave event data V2. The second filter unit 103b then discards the m 2 rows and n 2 columns of the first octave event data V2. The second filter unit 103b repeats such processing for all rows of the second octa
  • the second filter unit 103b outputs the third octave event data V3 of r 2 rows and s 2 columns to the third filter unit 103c and the event synthesis unit 104.
  • the second octave event data V3 generated by the second filter unit 103b is sent to the event synthesis unit 104, and a copy thereof is sent to the third filter unit 103c.
  • FIG. 3 shows the third octave event data V3 as a schematic representation of a plurality of regions P3 arranged in a row.
  • the process of generating the third octave event data V3 from the second octave event data V2 is repeated M ⁇ N/16 times in total.
  • four regions P2 are aggregated into one region P3, and therefore each region P3 corresponds to event data obtained by aggregating event data for 16 pixels 101a.
  • the third octave event data V3 is also called the third octave event ignition presence/absence column.
  • the operation of the third filter unit 103c is the same as that of the first filter unit 103a and the second filter unit 103b.
  • the third filter unit 103c generates the fourth octave event data V4 of r 3 rows and s 3 columns (r 3 is an integer satisfying r 3 ⁇ m 3 , s 3 is an integer satisfying s 3 ⁇ n 3 ) from the third octave event data V3 of m 3 rows and n 3 columns (n 3 is an integer satisfying 2 or more, n 3 is an integer satisfying 2 or more).
  • the third filter unit 103c outputs the fourth octave event data V4 of r 3 rows and s 3 columns to the event synthesis unit 104.
  • FIG. 3 shows the fourth octave event data V4 as a plurality of regions P4 arranged in a row.
  • the event data V4 of the fourth octave is also called the event firing presence/absence column of the fourth octave.
  • the event generating unit 103 generates event data Vj+1 of the j+1th octave from event data Vj of the jth octave (j is an integer satisfying 1 ⁇ j ⁇ i-1). This makes it possible to sequentially generate event data V2 to Vi of the second to i-th octaves from event data V1 of the first octave.
  • the term "octave” in this embodiment is used by analogy with the musical term "octave” because the differences between the event data V1 to Vi of the first to i-th octaves correspond to differences in frequency.
  • the event synthesis unit 104 acquires the event data V1 of the first octave from the event acquisition unit 102, and acquires the event data V2 to Vi of the second to i-th octaves from the event generation unit 103.
  • the octave information adding unit 104a adds octave information to the event data V1 to Vi of the first to i-th octaves and stores the added information.
  • the octave information is identification information for the event data V1 to Vi of the first to i-th octaves.
  • the octave information is based on the number of octaves of the event data, and is, for example, a value obtained by subtracting 1 from the number of octaves of the event data. Therefore, the octave information for the event data V1 to Vi of the first to i-th octaves is "0 to i-1", respectively.
  • the octave information for the event data V1 of the first octave is "0"
  • the octave information for the event data V2 of the second octave is "1”
  • the octave information for the event data V3 of the third octave is "2".
  • the output timing adjustment unit 104b adjusts the timing at which the event data held by the octave information addition unit 104a is output to the event output unit 105.
  • the output timing adjustment unit 104b outputs the event data V1 to Vi of the first to i-th octaves to the event output unit 105 in order from the event data with the largest octave number to the event data with the largest octave number.
  • the event data V1 to Vi of the first to i-th octaves are output with octave information added.
  • Event output unit 105 The event output unit 105 outputs at least a part of the event data V1 to Vi of the first to i-th octaves to the outside of the solid-state imaging device 100.
  • the event output unit 105 outputs the event data acquired from the event synthesis unit 104.
  • the event output unit 105 of the present embodiment outputs the event data to the vehicle control system 11 shown in Fig. 1, but may instead output the event data to the electronic device 200 shown in Fig. 10 or 11.
  • the event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves. For example, when the event data V1 to V3 of the first to third octaves are selected from the event data V1 to Vi of the first to i-th octaves, the event data V1 to V3 of the first to third octaves are output to the event data formation unit 105b.
  • the event data forming unit 105b converts the event data selected by the event data selecting unit 105a into an event output data format.
  • the event data forming unit 105b then outputs the event data converted into the event output data format to the outside of the solid-state imaging device 100 with the octave information added.
  • FIG. 4 is a diagram for explaining the pixel array 101 of the first embodiment.
  • Each of A to C in Figure 4 shows a pixel array 101 that outputs event data, and an image E obtained by converting this event data into an image representation.
  • Image E corresponds to an image captured when a situation in which the letter "A" newly appears.
  • Image E also corresponds to an image obtained from event data V1 of the first octave.
  • FIGS 4A to 4C represent the event data of each pixel 101a.
  • x0 and y0 represent the coordinates of each pixel 101a.
  • t0 represents the time when the event data was obtained.
  • p0 represents the polarity of the event data. For example, the polarity when an on-event is fired is "+", and the polarity when an off-event is fired is "-". The above also applies to the other x, y, t, and p.
  • a in FIG. 4 shows a pixel array 101 with a small number of pixels and low resolution.
  • B in FIG. 4 shows a pixel array 101 with a medium number of pixels and medium resolution.
  • C in FIG. 4 shows a pixel array 101 with a large number of pixels and high resolution.
  • the data mining costs will increase and the information processing of the information processing system that uses the event data output from the solid-state imaging device 100 will be delayed. As a result, the performance of the entire information processing system will actually decrease. Examples of such information processing systems are the vehicle control system 11 shown in FIG. 1, the electronic device 200 shown in FIG. 10, and the system shown in FIG. 11 (a system including the electronic device 200).
  • the solid-state imaging device 100 of this embodiment therefore outputs event data of various octaves, as described with reference to FIG. 3. This makes it possible to achieve high resolution for the solid-state imaging device 100 while suppressing delays in information processing using the event data. For example, even if the pixel array 101 shown in FIG. 4C is employed, it is possible to suppress delays in information processing using the event data. Further details of such effects will be described later.
  • FIG. 5 is a diagram for explaining the operation of the solid-state imaging device 100 of the first embodiment.
  • Arrow A1 in Figure 5 indicates the process of generating second octave event data V2 of one row and one column from first octave event data V1 of two rows and two columns. In this process, area P1 of two rows and two columns is replaced with area P2 of one row and one column.
  • Arrow A2 in Figure 5 indicates the process of generating third octave event data V3 of one row and one column from second octave event data V2 of two rows and two columns. In this process, area P2 of two rows and two columns is replaced with area P3 of one row and one column. The presence or absence of a check in areas P1 to P3 indicates whether an event has been fired or not.
  • FIG. 5 further shows an image E1 obtained by converting the event data V1 of the first octave into an image representation, an image E2 obtained by converting the event data V2 of the second octave into an image representation, and an image E3 obtained by converting the event data V3 of the third octave into an image representation.
  • the images E1, E2, and E3 are similar to the image E shown in C, B, and A of FIG. 4, respectively.
  • it is possible to generate a low-resolution image from a high-resolution image by generating event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave.
  • the event output unit 105 outputs the event data in an image representation for each octave.
  • the event data for each octave in a form such as images E1 to E3 by using the event data output in an image representation.
  • the event data is expressed in the form of (oc, x, y, t, p) by adding octave information.
  • x and y represent the coordinates of the pixels 101a corresponding to the regions P1 to P3, for example, the coordinates of one pixel 101a corresponding to the region P1, the average coordinates of four pixels 101a corresponding to the region P2, or the average coordinates of 16 pixels 101a corresponding to the region P3.
  • t represents the time when the event data was obtained.
  • p represents the polarity of the event data.
  • oc represents the octave information of the event data.
  • the octave information oc of the event data V1 in the first octave is "0"
  • the octave information oc of the event data V2 in the second octave is "1”
  • the octave information oc of the event data V3 in the third octave is "2".
  • FIG. 6 is a diagram for explaining the operation of the event generation unit 103 in the first embodiment.
  • a in FIG. 6 shows the process of generating r rows and s columns of second octave event data V2 from m rows and n columns of first octave event data V1.
  • 1 row and 1 column of j+1th octave event data Vj+1 is generated from 2 rows and 2 columns of jth octave event data Vj.
  • k represents the number of event firings contained in the m ⁇ n regions P1 (k is an integer satisfying 1 ⁇ k ⁇ m ⁇ n). For example, if the number of event firings contained in the 2 ⁇ 2 regions P1 is 3, the value of k is 3. In this case, three of the 2 ⁇ 2 regions P1 correspond to the "checked" region P1 shown in FIG. 5, and the remaining one of the 2 ⁇ 2 regions P1 corresponds to the "unchecked" region P1 shown in FIG. 5. This indicates that events were fired in three of the four pixels 101a, and no event was fired in the remaining one of the four pixels 101a.
  • FIG. 7 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
  • a to C in FIG. 7 correspond to A to C in FIG. 6, respectively.
  • one region P2 is "checked” (C in FIG. 7)
  • one region P2 is "unchecked” (B in FIG. 7).
  • an event is fired in two or more of the four regions P1
  • it is treated as if an event was fired in one region P2.
  • it is treated as if an event was not fired in one region P2.
  • the same is true when generating event data Vj+1 in the j+1th octave from event data Vj in the jth octave.
  • FIG. 8 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
  • a to C in FIG. 8 correspond to A to C in FIG. 6, respectively.
  • one region P2 is "checked” (C in FIG. 8)
  • one region P2 is "unchecked” (B in FIG. 8).
  • an event is fired in two or three of the four regions P1
  • it is treated as if an event was fired in one region P2.
  • it is treated as if an event was not fired in one region P2. This is also the case when generating event data Vj+1 in the j+1th octave from event data Vj in the jth octave.
  • FIG. 9 is a diagram for explaining the operation of the event output unit 105 in the first embodiment.
  • a in FIG. 9 is a diagram for explaining the operation of the event data selection unit 105a.
  • the event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves.
  • the event data selection unit 105a may select the event data V1 to Vi of all octaves, or may select only the event data V1 to Vi of some octaves.
  • the event data V1 to V3 of the first to third octaves are selected from the event data V1 to Vi of the first to i-th octaves
  • the event data V1 to V3 of the first to third octaves are output to the event data formation unit 105b.
  • the event data V1 to V3 of the first to third octaves are output with the octave information "0 to 2" added, respectively.
  • the event data formation unit 105b converts the event data selected by the event data selection unit 105a into an event output data format.
  • the event output unit 105 converts the event data selected by the event data selection unit 105a into an address event representation or an image representation for each octave.
  • the event data formation unit 105b then outputs the event data converted into the event output data format to the outside of the solid-state imaging device 100 with octave information added.
  • the event data output in the image representation can be used to display the event data for each octave in a format such as images E1 to E3.
  • FIG. 10 is a diagram illustrating an example of the electronic device 200 according to the first embodiment.
  • a in FIG. 10 shows a smartphone with a camera function as an example of electronic device 200.
  • this electronic device 200 includes an imaging unit 201, a display unit 202, an information processing unit 203, a storage unit 204, and an input unit 205.
  • the information processing unit 203 includes an extraction unit 203a and a selection unit 203b.
  • the imaging unit 201 is a functional block for implementing a camera function.
  • the imaging unit 201 includes the solid-state imaging device 100 shown in FIG. 3.
  • This electronic device 200 functions as an information processing system that performs information processing using event data output from the solid-state imaging device 100 (event output unit 105).
  • the display unit 202 has a display screen for displaying characters and images.
  • the display unit 202 displays event data output from the solid-state imaging device 100 on the display screen.
  • the event output unit 105 outputs the event data in an image representation
  • the display unit 202 displays this event data in the form of an image on the display screen (B in FIG. 10).
  • This image may be a still image or a moving image.
  • the display screen displays an image captured by the solid-state imaging device 100 as an image sensor, and an image in which the event data output from the solid-state imaging device 100 as an event sensor (EVS) is displayed in the form of an image.
  • EVS event sensor
  • the display screen In a mode in which event data is displayed on the display screen, the display screen is also called a viewer.
  • the information processing unit 203 performs various information processes such as controlling the electronic device 200. For example, the information processing unit 203 receives event data from the solid-state imaging device 100 and displays the event data on the display screen of the display unit 202.
  • the storage unit 204 includes a recording medium such as a semiconductor memory.
  • the information processing unit 203 can read information necessary for information processing from the storage unit 204, and record information generated by information processing in the storage unit 204.
  • the information processing unit 203 receives event data from the solid-state imaging device 100, and records this event data in the storage unit 204.
  • the input unit 205 accepts input operations from the user.
  • the information processing unit 203 performs information processing according to the input operations.
  • the input unit 205 includes, for example, a touch panel and hard buttons.
  • the event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves. For example, when the event data V1 to V3 of the first to third octaves is selected from the event data V1 to Vi of the first to i-th octaves, the event data V1 to V3 of the first to third octaves is output to the event data formation unit 105b.
  • the event data formation unit 105b outputs the event data selected by the event data selection unit 105a in an image representation for each octave. For example, the event data V1 to V3 of the first to third octaves is output in an image representation for each octave.
  • the extraction unit 203a extracts event data of a predetermined number of octaves from the event data output from the solid-state imaging device 100 (event data formation unit 105b). For example, the extraction unit 203a extracts event data V2 of the second octave from event data V1 to V3 of the first to third octaves. Event data of a predetermined number of octaves can be extracted based on the octave information of the event data.
  • the information processing unit 203 displays the event data extracted by the extraction unit 203a on the display screen. For example, when event data V2 of the second octave is extracted, the event data V2 of the second octave is displayed on the display screen in the form of an image.
  • the extraction unit 203a automatically extracts event data for the number of octaves that match the resolution of the viewer. For example, when event data is first displayed in the viewer, event data V2 of the second octave is extracted and displayed. Thereafter, when the user performs an operation to increase the resolution of the viewer, event data V1 of the first octave is extracted and displayed. On the other hand, when the user performs an operation to decrease the resolution of the viewer, event data V3 of the third octave is extracted and displayed.
  • the selection unit 203b selects the event data for the number of octaves specified by the user from the event data output from the solid-state imaging device 100 (event data formation unit 105b). For example, when the event data V2 for the second octave is displayed on the viewer, if the user touches the "recording start button" on the touch panel, the selection unit 203b selects the event data V2 for the second octave from the event data V1 to V3 for the first to third octaves. Then, the selection unit 203b starts recording the event data V2 for the second octave in the storage unit 204.
  • the recording of the event data V2 for the second octave ends.
  • the user can specify the number of octaves of the event data to be recorded and the timing to start and end recording.
  • a video from the start of recording to the end of recording is recorded in the storage unit 204 (recording medium).
  • the number of octaves of the event data recorded in the memory unit 204 may be different from the number of octaves of the event data displayed in the viewer. For example, when the event data V2 of the second octave is displayed in the viewer, if the user specifies "event data V1 of the first octave" as "to be recorded” on the touch panel, the event data V1 of the first octave may be recorded in the memory unit 204. Furthermore, even if no such specification is made, the event data V1 of the first octave may be recorded in the memory unit 204.
  • the information processing unit 203 may use the extracted event data for information processing other than display. For example, after the event data V2 of the second octave is extracted, the information processing unit 203 may use the event data V2 of the second octave for image recognition.
  • image recognition is recognition of a user's gesture.
  • the information processing unit 203 may perform image recognition to recognize a user's gesture using an image of the user included in the event data V2 of the second octave. At this time, the information processing unit 203 may use event data of multiple octaves for image recognition.
  • the solid-state imaging device 100 outputs event data of various octaves. Therefore, the information processing unit 203 can display event data of various octaves by extracting event data of a predetermined number of octaves from the output event data. In addition, the information processing unit 203 can record event data of various octaves by selecting event data of a predetermined number of octaves from the output event data. If the information processing unit 203 were to generate event data of a different resolution from event data of a certain resolution, the information processing by the information processing unit 203 would be delayed.
  • the solid-state imaging device 100 is responsible for the process of generating event data of various resolutions (number of octaves), so that the delay in information processing by the information processing unit 203 can be suppressed.
  • the process of generating event data of various resolutions is performed by the solid-state imaging device 100 as hardware, instead of by the information processing unit 203 as software.
  • FIG. 11 is a diagram showing another example of the electronic device 200 of the first embodiment.
  • a in FIG. 11 shows a game console as an example of electronic device 200.
  • this electronic device 200 is used by being connected to an imaging device 201' and a display device 202' via wired or wireless connection.
  • this electronic device 200 includes an information processing unit 203, a storage unit 204, and an input unit 205.
  • the information processing unit 203 includes an extraction unit 203a and a selection unit 203b.
  • the imaging device 201' is, for example, a camera that is an accessory to a game console.
  • the imaging device 201' includes the solid-state imaging device 100 shown in FIG. 3, similar to the imaging unit 201 described above.
  • This electronic device 200 together with the imaging device 201' and the display device 202', constitutes an information processing system that performs information processing using event data output from the solid-state imaging device 100 (event output unit 105).
  • the display device 202' is, for example, a large LCD television.
  • the display device 202' has a display screen for displaying characters and images, similar to the display unit 202 described above.
  • the imaging device 201' captures an image of a user playing on a game console.
  • the display device 202' captures an image of the user and displays the event data obtained in the form of an image on the display screen.
  • the display screen is also called a viewer.
  • the functions of the information processing unit 203, memory unit 204, and input unit 205 shown in FIG. 11B are generally similar to the functions of the information processing unit 203, memory unit 204, and input unit 205 shown in FIG. 10A.
  • FIG. 12 is a diagram for explaining the details of the electronic device 200 shown in FIG. 11.
  • the imaging device 201' captures the entire body of the user. Therefore, the event data V1 to V3 of the first to third octaves output from the imaging device 201' (solid-state imaging device 100) includes event data related to the entire body of the user.
  • the display screen of the display device 202' displays the user's entire body at low resolution using the third octave event data V3.
  • Image E3 in the area enclosed by the dotted line in FIG. 12A includes the user's entire body.
  • image E3 is obtained.
  • the display screen of the display device 202' displays the user's entire body at medium resolution using the second octave event data V2.
  • Image E2 in the area enclosed by the dotted line in FIG. 12B includes the user's hands. This image E2 corresponds to a portion of the second octave event data V2.
  • the display screen of the display device 202' displays the user's hand in high resolution using the event data V1 of the first octave.
  • Image E1 in the area enclosed by the dotted line in FIG. 12C includes the user's hand. This image E1 corresponds to a portion of the event data V1 of the first octave.
  • the information processing unit 203 can use these images E1 to E3 to enlarge or reduce the entire body of the user for display. For example, by transitioning the display content of the display screen from A in FIG. 12 to B in FIG. 12, the entire body of the user can be enlarged and displayed while increasing the image resolution. Also, by transitioning the display content of the display screen from B in FIG. 12 to C in FIG. 12, the hand of the user can be enlarged and displayed while increasing the image resolution. This makes it possible to confirm the gesture of the user's hand on the display screen. The transition from A in FIG. 12 to B in FIG.
  • the transition from B in FIG. 12 to C in FIG. 12 can be realized by switching the event data extracted by the extraction unit 203a from the event data V2 of the second octave to the event data V1 of the first octave.
  • the information processing unit 203 may perform image recognition using the event data extracted by the extraction unit 203a to automate the confirmation of gestures. For example, the information processing unit 203 extracts the area of image E3, i.e., the user's entire body, from the event data V3 of the third octave by image recognition. Next, the information processing unit 203 extracts the area of image E2, i.e., the user's hand part, from the event data V2 of the second octave by image recognition. Next, the information processing unit 203 identifies the user's hand gesture from the event data V1 of the first octave by image recognition. This makes it possible for the information processing unit 203 to automatically recognize the user's hand gesture. According to this embodiment, by having the solid-state imaging device 100 take charge of the process of generating event data of various resolutions (number of octaves), it is possible to suppress such delays in image recognition.
  • the solid-state imaging device 100 of this embodiment treats the event data output from the pixel array 101 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave. Therefore, according to this embodiment, by outputting event data of various octaves for information processing using the event data, it is possible to achieve high resolution of the solid-state imaging device 100 while suppressing delays in information processing.
  • FIG. 13 is a block diagram showing the configuration of a solid-state imaging device 100 according to the second embodiment.
  • the solid-state imaging device 100 of this embodiment like the solid-state imaging device of the first embodiment, includes a pixel array 101, an event acquisition unit 102, an event generation unit 103, an event synthesis unit 104, and an event output unit 105.
  • the solid-state imaging device 100 of this embodiment further includes a frame memory 111.
  • the event acquisition unit 102 of this embodiment also includes an arbiter unit 102a and a time stamp unit 102b.
  • the solid-state imaging device 100 of this embodiment is an arbiter-type EVS. Therefore, the event acquisition unit 102 of this embodiment acquires event data from multiple pixels 102a in the pixel array 102 in a random order.
  • the event data acquired by the event acquisition unit 102 is stored in the frame memory 111 for a certain period of time, and is then output from the frame memory 111 to the event generation unit 103 and the event synthesis unit 104.
  • the frame memory 111 includes multiple memory cells arranged in a two-dimensional array (matrix), similar to the pixel array 101.
  • the event data of each row of the frame memory 111 is treated as event data V1 of the first octave.
  • the event acquisition unit 102 of this embodiment includes an arbiter unit 102a and a time stamp unit 102b as functional blocks for an arbiter-type EVS.
  • the arbiter unit 102a arbitrates multiple events (request signals) output from multiple pixels 101a.
  • the time stamp unit 102b assigns a time stamp to an event fired from each pixel 101a.
  • the value of t in the event data (x, y, t, p) of each pixel 101a becomes the value of the time stamp.
  • FIG. 14 is a diagram for explaining the operation of the event generation unit 103 in the second embodiment.
  • FIG. 14 like A of FIG. 6, A of FIG. 7, and A of FIG. 8, shows the process of generating r rows and s columns of second octave event data V2 from m rows and n columns of first octave event data V1.
  • the processing in the example shown in FIG. 14 can be performed in the same manner as the processing in the example shown in FIG. 6A, FIG. 7A, or FIG. 8A.
  • the value of t in the event data (x, y, t, p) of each pixel 101a becomes the value of the timestamp, so the four regions P1 shown in FIG. 14 have different values of t.
  • the values of t in the four regions P1 are ta, tb, tc, and td, respectively.
  • the event data of this embodiment is expressed in the form (oc, x, y, t, p) with the addition of octave information.
  • x and y represent the coordinates of the pixels 101a corresponding to regions P1 to P3, etc., and represent, for example, the coordinates of one pixel 101a corresponding to region P1, the average coordinates of four pixels 101a corresponding to region P2, or the average coordinates of 16 pixels 101a corresponding to region P3.
  • t represents the time when the event data was obtained.
  • p represents the polarity of the event data.
  • the t of one region P2 is taken as the statistical value of the t of four regions P1.
  • the t of one region P2 is taken as the value obtained by performing statistical processing on the t of four regions P1.
  • the statistical value of the t of four regions P1 is, for example, the average value, maximum value, minimum value, etc. of the t of the four regions P1.
  • the t of one region Pj+1 is taken as the statistical value of the t of the four regions Pj.
  • the solid-state imaging device 100 of this embodiment treats the event data output from the frame memory 111 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave. Therefore, according to this embodiment, as in the first embodiment, by outputting event data of various octaves for information processing using the event data, it is possible to achieve high resolution of the solid-state imaging device 100 while suppressing delays in information processing.
  • FIG. 15 is a perspective view that illustrates a schematic configuration of a solid-state imaging device 100 according to the third embodiment.
  • the solid-state imaging device 100 of this embodiment includes a detection chip 120 and a light-receiving chip 130 stacked on the detection chip 120.
  • the detection chip 120 and the light-receiving chip 130 are electrically connected through connections such as via plugs, metal pads, and metal bumps.
  • the solid-state imaging device 100 of this embodiment functions as the solid-state imaging device 100 of the first or second embodiment.
  • FIG. 15 shows the X-axis, Y-axis, and Z-axis, which are perpendicular to each other.
  • the X-axis and Y-axis correspond to the horizontal direction
  • the Z-axis corresponds to the vertical direction.
  • the +Z direction corresponds to the upward direction
  • the -Z direction corresponds to the downward direction. Note that the -Z direction may or may not strictly coincide with the direction of gravity.
  • FIG. 16 is a plan view that shows a schematic configuration of the light receiving chip 120 of the third embodiment.
  • a in FIG. 16 shows an example of the planar structure of the light-receiving chip 120.
  • the light-receiving chip 120 includes a light-receiving section 121 and multiple via arrangement sections 122 to 124.
  • B in FIG. 16 shows an example of the planar structure of the light-receiving section 121.
  • the light-receiving section 121 includes multiple photodiodes 121a.
  • multiple photodiodes 121a are arranged in an array (two-dimensional lattice).
  • a pixel address consisting of a row address and a column address is assigned to each photodiode 121a, and each photodiode 121a is treated as a pixel.
  • Each photodiode 121a photoelectrically converts incident light to generate a photocurrent.
  • Via plugs electrically connected to the detection chip 130 are arranged in the via arrangement sections 122 to 124.
  • FIG. 17 is a plan view that shows a schematic configuration of the detection chip 130 of the third embodiment.
  • a in FIG. 17 shows an example of the planar structure of the detection chip 130.
  • the detection chip 130 includes an address event detection unit 131, multiple via placement units 132 to 134, a row driving circuit 135, a column driving circuit 136, and a signal processing circuit 137.
  • B in FIG. 17 shows an example of the planar structure of the address event detection unit 131.
  • the address event detection unit 131 includes multiple address event detection circuits 131a.
  • multiple address event detection circuits 131a are arranged in an array (two-dimensional lattice).
  • a pixel address is assigned to each address event detection circuit 131a, and each address event detection circuit 131a is electrically connected to the photodiode 121a at the same address.
  • Each address event detection circuit 121a quantizes a voltage signal corresponding to the photocurrent from the corresponding photodiode 121a and outputs it as a detection signal.
  • This detection signal is a one-bit signal indicating whether or not an address event has been detected in which the amount of incident light has exceeded a predetermined threshold, and is output to the signal processing circuit 137.
  • Via plugs electrically connected to the light receiving chip 120 are arranged in the via arrangement sections 132 to 134.
  • the row driving circuit 135 selects a row address and outputs a detection signal corresponding to that row address to the address event detection unit 131.
  • the column driving circuit 136 selects a column address and outputs a detection signal corresponding to that column address to the address event detection unit 131.
  • the signal processing circuit 137 performs predetermined signal processing on the detection signal from the address event detection unit 131.
  • the signal processing circuit 137 arranges the detection signals as pixel signals in a two-dimensional lattice pattern and acquires image data having one bit of information for each pixel.
  • the signal processing circuit 137 performs signal processing such as image recognition processing on this image data.
  • FIG. 18 is a circuit diagram showing the configuration of each address event detection circuit 131a in the third embodiment.
  • Each address event detection circuit 131a includes a current-voltage conversion circuit 310, a buffer 320, a subtractor 330, a quantizer 340, and a transfer circuit 350.
  • the current-voltage conversion circuit 310 converts the photocurrent from the corresponding photodiode 121a into a voltage signal.
  • the current-voltage conversion circuit 310 supplies this voltage signal to the buffer 320.
  • the buffer 320 corrects the voltage signal from the current-voltage conversion circuit 310.
  • the buffer 320 outputs the corrected voltage signal to the subtractor 330.
  • the subtractor 330 reduces the level of the voltage signal from the buffer 320 in accordance with the row drive signal from the row drive circuit 135.
  • the subtractor 330 supplies the reduced voltage signal to the quantizer 340.
  • the quantizer 340 quantizes the voltage signal from the subtractor 330 into a digital signal and outputs it as a detection signal.
  • the quantizer 340 outputs this detection signal to the transfer circuit 350.
  • the transfer circuit 350 transfers the detection signal from the quantizer 340 to the signal processing circuit 137 in accordance with the column drive signal from the column drive circuit 136.
  • FIG. 19 is a circuit diagram showing the configuration of a current-voltage conversion circuit 310 according to the third embodiment.
  • the current-voltage conversion circuit 310 includes an N-type transistor 311, a P-type transistor 312, and an N-type transistor 313.
  • These N-type and P-type transistors 311 to 313 are, for example, MOS (Metal-Oxide-Semiconductor) transistors.
  • the source of the N-type transistor 311 is electrically connected to the cathode of the photodiode 121a, and the drain of the N-type transistor 311 is electrically connected to the power supply terminal (VDD).
  • the P-type transistor 312 and the N-type transistor 313 are connected in series between the power supply terminal and the ground terminal (GND).
  • the node between the P-type transistor 312 and the N-type transistor 313 is electrically connected to the gate of the N-type transistor 311 and the input terminal of the buffer 320.
  • a predetermined bias voltage Vbias1 is applied to the gate of the P-type transistor 312.
  • the node between the N-type transistor 311 and the photodiode 121a is electrically connected to the gate of the N-type transistor 313.
  • N-type transistor 311 and the drain of N-type transistor 313 are placed on the power supply side, and this type of circuit is called a source follower.
  • the photocurrent from photodiode 121a is converted into a voltage signal by the source follower.
  • P-type transistor 312 supplies a constant current to N-type transistor 313. Note that the ground of the light receiving chip 120 and the ground of the detection chip 130 are separated from each other to prevent interference.
  • FIG. 20 is a circuit diagram showing the configuration of the subtractor 330 and quantizer 340 of the third embodiment.
  • the subtractor 330 includes a capacitor 331, an inverter 332, a capacitor 333, and a switch 334.
  • the quantizer 340 includes a comparator 341.
  • One electrode of the capacitor 331 is electrically connected to the output terminal of the buffer 320, and the other electrode of the capacitor 331 is electrically connected to the input terminal of the inverter 332.
  • the inverter 332 inverts the voltage signal input via the capacitor 331 and outputs the inverted signal to the non-inverting input terminal (+) of the comparator 341.
  • the capacitor 333 is connected in parallel to the inverter 332.
  • the switch 334 opens and closes the path that electrically connects both electrodes of the capacitor 333 in accordance with the row drive signal.
  • Equation 5 represents the subtraction operation of the voltage signal, and the gain of the subtraction result is C1/C2. Since it is usually desired to maximize the gain, it is preferable to set C1 large and design C2 small. On the other hand, if C2 is too small, kTC noise increases and noise characteristics may deteriorate, so the reduction in the capacitance of C2 is limited to a range in which noise can be tolerated. Furthermore, since an address event detection circuit 131a including a subtractor 330 is mounted for each pixel, there are area restrictions on C1 and C2. Taking these into consideration, for example, C1 is set to a value of 20 to 200 femtofarads (fF), and C2 is set to a value of 1 to 20 femtofarads (fF).
  • Comparator 341 compares the voltage signal from subtractor 330 with a predetermined threshold voltage Vth applied to the inverting input terminal (-). Comparator 341 outputs a signal indicating the comparison result as a detection signal to transfer circuit 350.
  • a simple pixel circuit including a photodiode and three to four transistors is provided for each pixel.
  • a complex pixel circuit including a photodiode 121a and an address event detection circuit 131a is provided for each pixel. Therefore, if both the photodiode 121a and the address event detection circuit 131a were placed on the same chip, the mounting area would be larger than in the case of the synchronous type. Therefore, in the solid-state imaging device 100 of this embodiment, the photodiode 121a and the address event detection circuit 131a are placed on the light receiving chip 120 and the detection chip 130, respectively. According to this embodiment, the mounting area can be reduced by distributing the photodiode 121a and the address event detection circuit 131a in this manner.
  • FIG. 21 is a circuit diagram showing the configuration of the light receiving chip 120 and the detection chip 130 of a modified example of the third embodiment.
  • N-type transistor 311, P-type transistor 312, and N-type transistor 313 in current-voltage conversion circuit 310 are arranged in detection chip 130.
  • N-type transistor 311 and N-type transistor 313 in current-voltage conversion circuit 310 are arranged in light-receiving chip 120
  • P-type transistor 312 in current-voltage conversion circuit 310 is arranged in detection chip 130.
  • the configuration in FIG. 21 is adopted, for example, when there is a risk that the circuit scale of detection chip 130 will increase with an increase in the number of pixels. According to this modified example, by arranging N-type transistor 311 and N-type transistor 313 in light-receiving chip 120, it is possible to reduce the circuit scale of detection chip 130.
  • N-type transistors 311, 313 were placed in the light receiving chip 120 and the other N-type transistor 311, 313 were placed in the detection chip 130, a process of forming the N-type transistor 120 in the light receiving chip 120 and a process of forming the N-type transistor 130 in the detection chip 130 would be required, increasing the number of processes for manufacturing the light receiving chip 120 and the detection chip 130.
  • this modified example by placing both N-type transistors 311, 313 in the light receiving chip 120, it is possible to reduce the number of processes for manufacturing the light receiving chip 120 and the detection chip 130. This makes it possible to reduce the manufacturing costs of the solid-state imaging device 100.
  • a solid-state imaging device comprising:
  • an octave information adding unit that adds octave information, which is identification information of the first to i-th octaves, to the event data of the first to i-th octaves, respectively;
  • the solid-state imaging device wherein the event output unit outputs the event data to which the octave information is added.
  • a frame memory for storing the event data output from the plurality of pixels, The solid-state imaging device according to (1), wherein the event generating unit generates the event data output from the frame memory as event data of the first octave.
  • An information processing system including a solid-state imaging device and an information processing unit
  • the solid-state imaging device includes: a plurality of pixels that detect an event and output event data indicative of the detection result of the event; an event generating unit that sets the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer equal to or greater than 2) from the event data of the first octave; an event output unit that outputs at least a part of the event data of the first to i-th octaves;
  • the information processing unit displays the event data output from the event output unit on a display screen.
  • the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, The information processing system according to (12), wherein the information processing unit displays the event data extracted by the extraction unit on the display screen.
  • the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, The information processing system according to (12), wherein the information processing unit performs image recognition using the event data extracted by the extraction unit.
  • the information processing unit includes a selection unit that selects the event data for a number of octaves designated by a user from the event data output from the event output unit, The information processing system according to (12), wherein the information processing unit records the event data selected by the selection unit on a recording medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Transforming Light Signals Into Electric Signals (AREA)
  • Traffic Control Systems (AREA)

Abstract

[Problem] To provide a solid-state imaging device and an information processing device which can achieve high resolution while suppressing delay in information processing. [Solution] A solid-state imaging device of the present disclosure comprises: a plurality of pixels that detect an event and output event data indicating detection results of the event; an event generation unit that establishes the event data output from the plurality of pixels as event data of a first octave, and generates second to i-th octave (i is an integer of two or greater) from the event data of the first octave; and an event output unit that outputs at least a portion of the event data of the first to i-th octave.

Description

固体撮像装置および情報処理システムSolid-state imaging device and information processing system
 本開示は、固体撮像装置および情報処理システムに関する。 This disclosure relates to a solid-state imaging device and an information processing system.
 イメージセンサが、被写体を含む画像を取得するためのセンサであるのに対し、イベントセンサ(EVS:Event-based Vision Sensor)は、被写体の変化を検出するためのセンサである。イベントセンサは、センシング対象を「画像」から「被写体の変化」に限定することで、イメージセンサに比べてフレームレートを高く設定することができる。イベントセンサは例えば、イメージセンサと同様に、CCD(Charge Coupled Device)センサやCMOS(Complementary Metal Oxide Semiconductor)センサなどの固体撮像装置により実現される。 While an image sensor is a sensor for acquiring images that include a subject, an event sensor (EVS: Event-based Vision Sensor) is a sensor for detecting changes in the subject. By limiting the sensing target from "images" to "changes in the subject," an event sensor can set a higher frame rate than an image sensor. Like an image sensor, an event sensor is realized by a solid-state imaging device such as a CCD (Charge Coupled Device) sensor or a CMOS (Complementary Metal Oxide Semiconductor) sensor.
特表2017-533497号公報JP 2017-533497 A
 従来のEVSは、一定の空間解像度のイベントデータを出力する。このイベントデータを受信した情報処理システムが物体認識などの情報処理を行う場合には、イベントデータに対してスケーリング、フィルタリング、クロッピングなどの処理を行いつつ、イベントデータから有用な情報を抽出していく。 Conventional EVSs output event data with a certain spatial resolution. When an information processing system receives this event data and performs information processing such as object recognition, it performs processes such as scaling, filtering, and cropping on the event data to extract useful information from it.
 しかしながら、EVSの性能を高めようとしてEVSを高解像度化すると、データマイニングコストが増大し、情報処理システムの情報処理が遅延してしまう。その結果、情報処理システム全体の性能はむしろ低下してしまう。 However, increasing the resolution of an EVS in an attempt to improve its performance increases data mining costs and slows down the information processing system's information processing. As a result, the performance of the entire information processing system actually decreases.
 例えば、イベントのアドレス情報の下位ビットをドロップすることで、イベントデータを空間的にダウンサンプリングするという手法が考えられている。これにより、イベントデータの空間解像度が低下するため、データマイニングコストの増大を抑制し、情報処理システムの情報処理の遅延を抑制することができる。しかしながら、空間解像度の低下により、物体認識などの目的タスクの精度が低下してしまうおそれがある。 For example, one method that has been considered is to spatially downsample event data by dropping the least significant bits of the event address information. This reduces the spatial resolution of the event data, thereby preventing increases in data mining costs and reducing delays in information processing in information processing systems. However, the reduction in spatial resolution may result in a decrease in the accuracy of target tasks such as object recognition.
 そこで、本開示は、情報処理の遅延を抑制しつつ、高解像度化を実現することが可能な固体撮像装置および情報処理システムを提供する。 The present disclosure provides a solid-state imaging device and information processing system that can achieve high resolution while suppressing delays in information processing.
 本開示の第1の側面の固体撮像装置は、イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部とを備える。これにより例えば、イベントデータを利用した情報処理用に様々なオクターブのイベントデータを出力することで、情報処理の遅延を抑制しつつ、固体撮像装置の高解像度化を実現することが可能となる。 The solid-state imaging device according to the first aspect of the present disclosure includes a plurality of pixels that detect an event and output event data indicating the detection result of the event, an event generation unit that converts the event data output from the plurality of pixels into event data of a first octave and generates event data of a second to i-th octave (i is an integer of 2 or more) from the event data of the first octave, and an event output unit that outputs at least a portion of the event data of the first to i-th octave. This makes it possible to achieve high resolution of the solid-state imaging device while suppressing delays in information processing, for example, by outputting event data of various octaves for information processing using the event data.
 また、この第1の側面の固体撮像装置は、前記第1~第iオクターブのイベントデータにそれぞれ、前記第1~第iオクターブの識別情報であるオクターブ情報を付加するオクターブ情報付加部をさらに備え、前記イベント出力部は、前記オクターブ情報が付加された前記イベントデータを出力してもよい。これにより例えば、あるイベントデータがどのオクターブのイベントデータかを識別することが可能となる。 The solid-state imaging device of the first aspect may further include an octave information adding unit that adds octave information, which is identification information for the first to i-th octaves, to the event data of the first to i-th octaves, respectively, and the event output unit may output the event data to which the octave information has been added. This makes it possible, for example, to identify which octave a certain event data belongs to.
 また、この第1の側面において、前記イベント出力部は、前記イベントデータをオクターブごとに画像表現で出力してもよい。これにより例えば、様々なオクターブのイベントデータを画像として表示することが可能となる。 Furthermore, in this first aspect, the event output unit may output the event data in an image representation for each octave. This makes it possible to display event data for various octaves as images, for example.
 また、この第1の側面において、前記イベント生成部は、第jオクターブ(jは1≦j≦i-1を満たす整数)のイベントデータから第j+1オクターブのイベントデータを生成してもよい。これにより例えば、第1オクターブのイベントデータから第2オクターブのイベントデータを生成し、第2オクターブのイベントデータから第3オクターブのイベントデータを生成するといったように、様々なオクターブのイベントデータを第1オクターブのイベントデータから順次生成することが可能となる。 In addition, in this first aspect, the event generating unit may generate event data for the j+1th octave from event data for the jth octave (j is an integer satisfying 1≦j≦i−1). This makes it possible to sequentially generate event data for various octaves from event data for the first octave, for example, by generating event data for the second octave from event data for the first octave, and generating event data for the third octave from event data for the second octave.
 また、この第1の側面において、前記イベント生成部は、m行分(mは2以上の整数)の前記第jオクターブのイベントデータから、1行分の前記第j+1オクターブのイベントデータを生成してもよい。これにより例えば、第jオクターブのイベントデータから第j+1オクターブのイベントデータを簡単に生成することが可能となる。 In addition, in this first aspect, the event generation unit may generate one line of event data for the j+1 octave from m lines (m is an integer equal to or greater than 2) of event data for the j octave. This makes it possible to easily generate event data for the j+1 octave from event data for the j octave, for example.
 また、この第1の側面において、前記イベント生成部は、m行n列分(nは2以上の整数)の前記第jオクターブのイベントデータから、1行1列分の前記第j+1オクターブのイベントデータを生成してもよい。これにより例えば、第jオクターブのイベントデータから第j+1オクターブのイベントデータをさらに簡単に生成することが可能となる。 In addition, in this first aspect, the event generation unit may generate one row and one column of event data for the j+1 octave from m rows and n columns (n is an integer of 2 or more) of event data for the j octave. This makes it possible to more easily generate event data for the j+1 octave from event data for the j octave, for example.
 また、この第1の側面において、前記イベント生成部は、前記m行n列分の前記第jオクターブのイベントデータがk個のイベント発火(kは1≦k≦m×nを満たす整数)を含む場合に、前記1行1列分の前記第j+1オクターブのイベントデータにてイベントを発火させてもよい。これにより例えば、第jオクターブのm×n個の領域を第j+1オクターブの1個の領域に簡単に置換することが可能となる。 In addition, in this first aspect, the event generating unit may ignite an event with the event data of the j+1 octave of the 1 row and 1 column when the event data of the j octave of the m rows and n columns includes k event firings (k is an integer satisfying 1≦k≦m×n). This makes it possible to easily replace, for example, m×n areas of the j th octave with one area of the j+1 th octave.
 また、この第1の側面において、mは2であり、nは2であり、かつkは1、2、3、および4でもよい。これにより例えば、第jオクターブのm×n個の領域でイベントが1個も発火しなかった場合と、第jオクターブのm×n個の領域でイベントが少なくとも1個発火した場合とを区別することが可能となる。 In addition, in this first aspect, m may be 2, n may be 2, and k may be 1, 2, 3, and 4. This makes it possible to distinguish, for example, between a case where no events are fired in the m×n regions of the jth octave and a case where at least one event is fired in the m×n regions of the jth octave.
 また、この第1の側面において、mは2であり、nは2であり、かつkは2、3、および4でもよい。これにより例えば、ノイズイベントの影響を抑制することが可能となる。 Also, in this first aspect, m may be 2, n may be 2, and k may be 2, 3, or 4. This makes it possible to suppress the effects of noise events, for example.
 また、この第1の側面において、mは2であり、nは2であり、かつkは3および4でもよい。これにより例えば、ノイズイベントおよびフリッカーイベントの影響を抑制することが可能となる。 In addition, in this first aspect, m may be 2, n may be 2, and k may be 3 and 4. This makes it possible to suppress the effects of noise events and flicker events, for example.
 また、この第1の側面の固体撮像装置は、前記複数の画素から出力された前記イベントデータを格納するフレームメモリをさらに備え、前記イベント生成部は、前記フレームメモリから出力された前記イベントデータを前記第1オクターブのイベントデータとしてもよい。これにより例えば、固体撮像装置としてアービタ型のイベントセンサを採用することが可能となる。 The solid-state imaging device of the first aspect may further include a frame memory that stores the event data output from the plurality of pixels, and the event generating unit may treat the event data output from the frame memory as event data of the first octave. This makes it possible to employ, for example, an arbiter-type event sensor as the solid-state imaging device.
 本開示の第2の側面の情報処理システムは、固体撮像装置と情報処理部とを備える情報処理システムであって、前記固体撮像装置は、イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部とを備え、前記情報処理部は、前記イベント出力部から出力された前記イベントデータを表示画面に表示する。これにより例えば、イベントデータを利用した情報処理(例えば表示)用に様々なオクターブのイベントデータを出力することで、情報処理の遅延を抑制しつつ、固体撮像装置の高解像度化を実現することが可能となる。 The information processing system according to a second aspect of the present disclosure is an information processing system including a solid-state imaging device and an information processing unit, in which the solid-state imaging device includes a plurality of pixels that detect an event and output event data indicating the detection result of the event, an event generation unit that treats the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer of 2 or more) from the event data of the first octave, and an event output unit that outputs at least a portion of the event data of the first to i-th octave, and the information processing unit displays the event data output from the event output unit on a display screen. As a result, for example, by outputting event data of various octaves for information processing (e.g., display) using the event data, it is possible to achieve high resolution of the solid-state imaging device while suppressing delays in information processing.
 また、この第2の側面において、前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、前記情報処理部は、前記抽出部により抽出された前記イベントデータを前記表示画面に表示してもよい。これにより例えば、固体撮像装置から出力された様々なオクターブのイベントデータを、オクターブごとに表示用に利用することが可能となる。 Also, in this second aspect, the information processing unit may include an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, and the information processing unit may display the event data extracted by the extraction unit on the display screen. This makes it possible, for example, to use the event data of various octaves output from a solid-state imaging device for display on an octave basis.
 また、この第2の側面において、前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、前記情報処理部は、前記抽出部により抽出された前記イベントデータを利用して画像認識を行ってもよい。これにより例えば、固体撮像装置から出力された様々なオクターブのイベントデータを、オクターブごとに画像認識用に利用することが可能となる。 Furthermore, in this second aspect, the information processing unit may include an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit, and the information processing unit may perform image recognition using the event data extracted by the extraction unit. This makes it possible, for example, to use the event data of various octaves output from a solid-state imaging device for image recognition on an octave-by-octave basis.
 また、この第2の側面において、前記画像認識は、ユーザーのジェスチャー認識でもよい。これにより例えば、固体撮像装置から出力された様々なオクターブのイベントデータを、オクターブごとにジェスチャー認識用に利用することが可能となる。 In addition, in this second aspect, the image recognition may be user gesture recognition. This makes it possible, for example, to use event data of various octaves output from a solid-state imaging device for gesture recognition for each octave.
 また、この第2の側面において、前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、ユーザーにより指定されたオクターブ数の前記イベントデータを選択する選択部を含み、前記情報処理部は、前記選択部により選択された前記イベントデータを記録媒体に記録してもよい。これにより例えば、固体撮像装置から出力された様々なオクターブのイベントデータを、オクターブごとに記録することが可能となる。 In addition, in this second aspect, the information processing unit may include a selection unit that selects the event data for a number of octaves specified by a user from the event data output from the event output unit, and the information processing unit may record the event data selected by the selection unit on a recording medium. This makes it possible to record, for example, event data for various octaves output from a solid-state imaging device for each octave.
 また、この第2の側面において、前記情報処理システムは、前記固体撮像装置と前記情報処理部とを備える電子機器でもよい。これにより例えば、イベントデータの出力から、その後の情報処理までを、同じ電子機器で行うことが可能となる。 In addition, in this second aspect, the information processing system may be an electronic device including the solid-state imaging device and the information processing unit. This makes it possible, for example, to perform the output of event data and the subsequent information processing in the same electronic device.
 また、この第2の側面において、前記電子機器はさらに、前記表示画面を有する表示部を備えてもいてよい。これにより例えば、イベントデータの出力から、イベントデータの表示までを、同じ電子機器で行うことが可能となる。 In addition, in this second aspect, the electronic device may further include a display unit having the display screen. This makes it possible, for example, to output the event data and display the event data on the same electronic device.
 また、この第2の側面において、前記情報処理システムは、前記情報処理部を含む電子機器と、前記電子機器の外部に設けられ、前記固体撮像装置を含む撮像装置とを備えていてもよい。これにより例えば、イベントデータを利用した情報処理を、撮像装置の外部の電子機器で行うことが可能となる。 In addition, in this second aspect, the information processing system may include an electronic device including the information processing unit, and an imaging device that is provided outside the electronic device and includes the solid-state imaging device. This makes it possible, for example, to perform information processing using event data in an electronic device outside the imaging device.
 また、この第2の側面において、前記情報処理システムはさらに、前記電子機器の外部に設けられ、前記表示画面を有する表示装置を備えていてもよい。これにより例えば、イベントデータの表示を、撮像装置の外部の電子機器で行うことが可能となる。 In addition, in this second aspect, the information processing system may further include a display device provided outside the electronic device and having the display screen. This makes it possible, for example, to display event data on an electronic device external to the imaging device.
第1実施形態の車両1の構成を示すブロック図である。1 is a block diagram showing a configuration of a vehicle 1 according to a first embodiment. 第1実施形態の車両1のセンシング領域を示す平面図である。FIG. 2 is a plan view showing a sensing area of the vehicle 1 according to the first embodiment. 第1実施形態の固体撮像装置100の構成を示すブロック図である。1 is a block diagram showing a configuration of a solid-state imaging device 100 according to a first embodiment. 第1実施形態の画素アレイ101について説明するための図である。2 is a diagram for explaining a pixel array 101 according to the first embodiment. FIG. 第1実施形態の固体撮像装置100の動作について説明するための図である。4A to 4C are diagrams for explaining the operation of the solid-state imaging device 100 according to the first embodiment. 第1実施形態のイベント生成部103の動作を説明するための図である。4 is a diagram for explaining the operation of an event generating unit 103 in the first embodiment. FIG. 第1実施形態のイベント生成部103の動作を説明するための別の図である。FIG. 11 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment. 第1実施形態のイベント生成部103の動作を説明するための別の図である。FIG. 11 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment. 第1実施形態のイベント出力部105の動作を説明するための図である。4A to 4C are diagrams for explaining the operation of an event output unit 105 according to the first embodiment. 第1実施形態の電子機器200の一例を示す図である。FIG. 1 illustrates an example of an electronic device 200 according to a first embodiment. 第1実施形態の電子機器200の別の一例を示す図である。FIG. 2 is a diagram illustrating another example of the electronic device 200 according to the first embodiment. 図11に示す電子機器200の詳細を説明するための図である。FIG. 12 is a diagram for explaining details of the electronic device 200 shown in FIG. 11 . 第2実施形態の固体撮像装置100の構成を示すブロック図である。FIG. 11 is a block diagram showing a configuration of a solid-state imaging device 100 according to a second embodiment. 第2実施形態のイベント生成部103の動作を説明するための図である。FIG. 11 is a diagram for explaining the operation of an event generating unit 103 in the second embodiment. 第3実施形態の固体撮像装置100の構成を模式的に示す斜視図である。FIG. 11 is a perspective view illustrating a schematic configuration of a solid-state imaging device 100 according to a third embodiment. 第3実施形態の受光チップ120の構成を模式的に示す平面図である。FIG. 13 is a plan view illustrating a schematic configuration of a photosensor chip 120 according to a third embodiment. 第3実施形態の検出チップ130の構成を模式的に示す平面図である。FIG. 13 is a plan view illustrating a schematic configuration of a detection chip 130 according to a third embodiment. 第3実施形態の各アドレスイベント検出回路131aの構成を示す回路図である。FIG. 13 is a circuit diagram showing a configuration of each address event detection circuit 131a according to a third embodiment. 第3実施形態の電流電圧変換回路310の構成を示す回路図である。FIG. 11 is a circuit diagram showing a configuration of a current-voltage conversion circuit 310 according to a third embodiment. 第3実施形態の減算器330と量子化器340の構成を示す回路図である。FIG. 11 is a circuit diagram showing configurations of a subtractor 330 and a quantizer 340 according to a third embodiment. 第3実施形態の変形例の受光チップ120と検出チップ130の構成を示す回路図である。FIG. 13 is a circuit diagram showing the configurations of the light receiving chip 120 and the detection chip 130 according to a modified example of the third embodiment.
 以下、本開示の実施形態を、図面を参照して説明する。 Embodiments of the present disclosure will be described below with reference to the drawings.
 (第1実施形態)
 (1)第1実施形態の車両1
 図1は、第1実施形態の車両1の構成を示すブロック図である。図1は、移動装置制御システムの一例である車両制御システム11の構成例を示している。
First Embodiment
(1) Vehicle 1 of the First Embodiment
Fig. 1 is a block diagram showing the configuration of a vehicle 1 according to the first embodiment. Fig. 1 shows an example of the configuration of a vehicle control system 11, which is an example of a mobility device control system.
 車両制御システム11は、車両1に設けられ、車両1の走行支援及び自動運転に関わる処理を行う。 The vehicle control system 11 is installed in the vehicle 1 and performs processing related to driving assistance and autonomous driving of the vehicle 1.
 車両制御システム11は、車両制御ECU(Electronic Control Unit)21、通信部22、地図情報蓄積部23、位置情報取得部24、外部認識センサ25、車内センサ26、車両センサ27、記憶部31、走行支援・自動運転制御部32、DMS(Driver Monitoring System)33、HMI(Human Machine Interface)34、及び、車両制御部35を備える。 The vehicle control system 11 includes a vehicle control ECU (Electronic Control Unit) 21, a communication unit 22, a map information storage unit 23, a location information acquisition unit 24, an external recognition sensor 25, an in-vehicle sensor 26, a vehicle sensor 27, a memory unit 31, a driving assistance/automated driving control unit 32, a DMS (Driver Monitoring System) 33, an HMI (Human Machine Interface) 34, and a vehicle control unit 35.
 車両制御ECU21、通信部22、地図情報蓄積部23、位置情報取得部24、外部認識センサ25、車内センサ26、車両センサ27、記憶部31、走行支援・自動運転制御部32、ドライバモニタリングシステム(DMS)33、ヒューマンマシーンインタフェース(HMI)34、及び、車両制御部35は、通信ネットワーク41を介して相互に通信可能に接続されている。通信ネットワーク41は、例えば、CAN(Controller Area Network)、LIN(Local Interconnect Network)、LAN(Local Area Network)、FlexRay(登録商標)、イーサネット(登録商標)といったディジタル双方向通信の規格に準拠した車載通信ネットワークやバス等により構成される。通信ネットワーク41は、伝送されるデータの種類によって使い分けられてもよい。例えば、車両制御に関するデータに対してCANが適用され、大容量データに対してイーサネットが適用されるようにしてもよい。なお、車両制御システム11の各部は、通信ネットワーク41を介さずに、例えば近距離無線通信(NFC(Near Field Communication))やBluetooth(登録商標)といった比較的近距離での通信を想定した無線通信を用いて直接的に接続される場合もある。 The vehicle control ECU 21, communication unit 22, map information storage unit 23, position information acquisition unit 24, external recognition sensor 25, in-vehicle sensor 26, vehicle sensor 27, memory unit 31, driving assistance/automatic driving control unit 32, driver monitoring system (DMS) 33, human machine interface (HMI) 34, and vehicle control unit 35 are connected to each other so as to be able to communicate with each other via a communication network 41. The communication network 41 is composed of an in-vehicle communication network or bus that complies with a digital two-way communication standard such as CAN (Controller Area Network), LIN (Local Interconnect Network), LAN (Local Area Network), FlexRay (registered trademark), or Ethernet (registered trademark). The communication network 41 may be used differently depending on the type of data being transmitted. For example, CAN may be applied to data related to vehicle control, and Ethernet may be applied to large-volume data. In addition, each part of the vehicle control system 11 may be directly connected without going through the communication network 41, using wireless communication intended for communication over relatively short distances, such as near field communication (NFC) or Bluetooth (registered trademark).
 なお、以下、車両制御システム11の各部が、通信ネットワーク41を介して通信を行う場合、通信ネットワーク41の記載を省略するものとする。例えば、車両制御ECU21と通信部22が通信ネットワーク41を介して通信を行う場合、単に車両制御ECU21と通信部22とが通信を行うと記載する。 Note that, hereinafter, when each part of the vehicle control system 11 communicates via the communication network 41, the description of the communication network 41 will be omitted. For example, when the vehicle control ECU 21 and the communication unit 22 communicate via the communication network 41, it will simply be described as the vehicle control ECU 21 and the communication unit 22 communicating with each other.
 [車両制御ECU21]
 車両制御ECU21は、例えば、CPU(Central Processing Unit)、MPU(Micro Processing Unit)といった各種のプロセッサにより構成される。車両制御ECU21は、車両制御システム11全体又は一部の機能の制御を行う。
[Vehicle control ECU 21]
The vehicle control ECU 21 is configured with various processors such as a CPU (Central Processing Unit), an MPU (Micro Processing Unit), etc. The vehicle control ECU 21 controls the entire or part of the functions of the vehicle control system 11.
 [通信部22]
 通信部22は、車内及び車外の様々な機器、他の車両、サーバ、基地局等と通信を行い、各種のデータの送受信を行う。このとき、通信部22は、複数の通信方式を用いて通信を行うことができる。
[Communication unit 22]
The communication unit 22 communicates with various devices inside and outside the vehicle, other vehicles, servers, base stations, etc., and transmits and receives various types of data. At this time, the communication unit 22 can communicate using a plurality of communication methods.
 通信部22が実行可能な車外との通信について、概略的に説明する。通信部22は、例えば、5G(第5世代移動通信システム)、LTE(Long Term Evolution)、DSRC(Dedicated Short Range Communications)等の無線通信方式により、基地局又はアクセスポイントを介して、外部ネットワーク上に存在するサーバ(以下、外部のサーバと呼ぶ)等と通信を行う。通信部22が通信を行う外部ネットワークは、例えば、インターネット、クラウドネットワーク、又は、事業者固有のネットワーク等である。通信部22が外部ネットワークに対して行う通信方式は、所定以上の通信速度、且つ、所定以上の距離間でディジタル双方向通信が可能な無線通信方式であれば、特に限定されない。 The following provides an overview of the communications with the outside of the vehicle that can be performed by the communication unit 22. The communication unit 22 communicates with servers (hereinafter referred to as external servers) on an external network via base stations or access points using wireless communication methods such as 5G (fifth generation mobile communication system), LTE (Long Term Evolution), and DSRC (Dedicated Short Range Communications). The external network with which the communication unit 22 communicates is, for example, the Internet, a cloud network, or an operator-specific network. The communication method that the communication unit 22 uses with the external network is not particularly limited as long as it is a wireless communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined distance.
 また例えば、通信部22は、P2P(Peer To Peer)技術を用いて、自車の近傍に存在する端末と通信を行うことができる。自車の近傍に存在する端末は、例えば、歩行者や自転車等の比較的低速で移動する移動体が装着する端末、店舗等に位置が固定されて設置される端末、又は、MTC(Machine Type Communication)端末である。さらに、通信部22は、V2X通信を行うこともできる。V2X通信とは、例えば、他の車両との間の車車間(Vehicle to Vehicle)通信、路側器等との間の路車間(Vehicle to Infrastructure)通信、家との間(Vehicle to Home)の通信、及び、歩行者が所持する端末等との間の歩車間(Vehicle to Pedestrian)通信等の、自車と他との通信をいう。 Furthermore, for example, the communication unit 22 can communicate with a terminal present in the vicinity of the vehicle using P2P (Peer To Peer) technology. The terminal present in the vicinity of the vehicle can be, for example, a terminal attached to a mobile object moving at a relatively slow speed, such as a pedestrian or a bicycle, a terminal installed at a fixed position in a store, or an MTC (Machine Type Communication) terminal. Furthermore, the communication unit 22 can also perform V2X communication. V2X communication refers to communication between the vehicle and others, such as vehicle-to-vehicle communication with other vehicles, vehicle-to-infrastructure communication with roadside devices, vehicle-to-home communication with a home, and vehicle-to-pedestrian communication with a terminal carried by a pedestrian, etc.
 通信部22は、例えば、車両制御システム11の動作を制御するソフトウエアを更新するためのプログラムを外部から受信することができる(Over The Air)。通信部22は、さらに、地図情報、交通情報、車両1の周囲の情報等を外部から受信することができる。また例えば、通信部22は、車両1に関する情報や、車両1の周囲の情報等を外部に送信することができる。通信部22が外部に送信する車両1に関する情報としては、例えば、車両1の状態を示すデータ、認識部73による認識結果等がある。さらに例えば、通信部22は、eコール等の車両緊急通報システムに対応した通信を行う。 The communication unit 22 can, for example, receive from the outside a program for updating the software that controls the operation of the vehicle control system 11 (Over the Air). The communication unit 22 can further receive map information, traffic information, information about the surroundings of the vehicle 1, etc. from the outside. For example, the communication unit 22 can also transmit information about the vehicle 1 and information about the surroundings of the vehicle 1 to the outside. Information about the vehicle 1 that the communication unit 22 transmits to the outside includes, for example, data indicating the state of the vehicle 1, the recognition results by the recognition unit 73, etc. Furthermore, for example, the communication unit 22 performs communication corresponding to a vehicle emergency notification system such as e-Call.
 例えば、通信部22は、電波ビーコン、光ビーコン、FM多重放送等の道路交通情報通信システム(VICS(Vehicle Information and Communication System)(登録商標))により送信される電磁波を受信する。 For example, the communication unit 22 receives electromagnetic waves transmitted by a road traffic information and communication system (VICS (Vehicle Information and Communication System) (registered trademark)) such as a radio beacon, optical beacon, or FM multiplex broadcasting.
 通信部22が実行可能な車内との通信について、概略的に説明する。通信部22は、例えば無線通信を用いて、車内の各機器と通信を行うことができる。通信部22は、例えば、無線LAN、Bluetooth、NFC、WUSB(Wireless USB)といった、無線通信により所定以上の通信速度でディジタル双方向通信が可能な通信方式により、車内の機器と無線通信を行うことができる。これに限らず、通信部22は、有線通信を用いて車内の各機器と通信を行うこともできる。例えば、通信部22は、図示しない接続端子に接続されるケーブルを介した有線通信により、車内の各機器と通信を行うことができる。通信部22は、例えば、USB(Universal Serial Bus)、HDMI(High-Definition Multimedia Interface)(登録商標)、MHL(Mobile High-definition Link)といった、有線通信により所定以上の通信速度でディジタル双方向通信が可能な通信方式により、車内の各機器と通信を行うことができる。 The following provides an overview of the communication that the communication unit 22 can perform with the inside of the vehicle. The communication unit 22 can communicate with each device in the vehicle using, for example, wireless communication. The communication unit 22 can perform wireless communication with each device in the vehicle using a communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined speed via wireless communication, such as wireless LAN, Bluetooth, NFC, or WUSB (Wireless USB). Not limited to this, the communication unit 22 can also communicate with each device in the vehicle using wired communication. For example, the communication unit 22 can communicate with each device in the vehicle using wired communication via a cable connected to a connection terminal (not shown). The communication unit 22 can communicate with each device in the vehicle using a communication method that allows digital two-way communication at a communication speed equal to or higher than a predetermined speed via wired communication, such as USB (Universal Serial Bus), HDMI (High-Definition Multimedia Interface) (registered trademark), or MHL (Mobile High-definition Link).
 ここで、車内の機器とは、例えば、車内において通信ネットワーク41に接続されていない機器を指す。車内の機器としては、例えば、運転者等の搭乗者が所持するモバイル機器やウェアラブル機器、車内に持ち込まれ一時的に設置される情報機器等が想定される。 Here, the term "devices in the vehicle" refers to devices that are not connected to the communication network 41 in the vehicle. Examples of devices in the vehicle include mobile devices and wearable devices carried by passengers such as the driver, and information devices that are brought into the vehicle and temporarily installed.
 [地図情報蓄積部23]
 地図情報蓄積部23は、外部から取得した地図及び車両1で作成した地図の一方又は両方を蓄積する。例えば、地図情報蓄積部23は、3次元の高精度地図、高精度地図より精度が低く、広いエリアをカバーするグローバルマップ等を蓄積する。
[Map information storage unit 23]
The map information storage unit 23 stores one or both of a map acquired from an external source and a map created by the vehicle 1. For example, the map information storage unit 23 stores a three-dimensional high-precision map, a global map that has lower precision than a high-precision map and covers a wide area, and the like.
 高精度地図は、例えば、ダイナミックマップ、ポイントクラウドマップ、ベクターマップ等である。ダイナミックマップは、例えば、動的情報、準動的情報、準静的情報、静的情報の4層からなる地図であり、外部のサーバ等から車両1に提供される。ポイントクラウドマップは、ポイントクラウド(点群データ)により構成される地図である。ベクターマップは、例えば、車線や信号機の位置といった交通情報等をポイントクラウドマップに対応付け、ADAS(Advanced Driver Assistance System)やAD(Autonomous Driving)に適合させた地図である。 High-precision maps include, for example, dynamic maps, point cloud maps, and vector maps. A dynamic map is, for example, a map consisting of four layers of dynamic information, semi-dynamic information, semi-static information, and static information, and is provided to the vehicle 1 from an external server or the like. A point cloud map is a map composed of a point cloud (point group data). A vector map is, for example, a map that associates traffic information such as the positions of lanes and traffic lights with a point cloud map, and is adapted for ADAS (Advanced Driver Assistance System) and AD (Autonomous Driving).
 ポイントクラウドマップ及びベクターマップは、例えば、外部のサーバ等から提供されてもよいし、カメラ51、レーダ52、LiDAR53等によるセンシング結果に基づいて、後述するローカルマップとのマッチングを行うための地図として車両1で作成され、地図情報蓄積部23に蓄積されてもよい。また、外部のサーバ等から高精度地図が提供される場合、通信容量を削減するため、車両1がこれから走行する計画経路に関する、例えば数百メートル四方の地図データが外部のサーバ等から取得される。 The point cloud map and vector map may be provided, for example, from an external server, or may be created by the vehicle 1 based on sensing results from the camera 51, radar 52, LiDAR 53, etc. as a map for matching with a local map described below, and stored in the map information storage unit 23. In addition, when a high-precision map is provided from an external server, etc., map data of, for example, an area of several hundred meters square regarding the planned route along which the vehicle 1 will travel is acquired from the external server, etc., in order to reduce communication capacity.
 [位置情報取得部24]
 位置情報取得部24は、GNSS(Global Navigation Satellite System)衛星からGNSS信号を受信し、車両1の位置情報を取得する。取得した位置情報は、走行支援・自動運転制御部32に供給される。なお、位置情報取得部24は、GNSS信号を用いた方式に限定されず、例えば、ビーコンを用いて位置情報を取得してもよい。
[Location information acquisition unit 24]
The position information acquisition unit 24 receives GNSS signals from Global Navigation Satellite System (GNSS) satellites and acquires position information of the vehicle 1. The acquired position information is supplied to the driving assistance/automated driving control unit 32. Note that the position information acquisition unit 24 is not limited to a method using GNSS signals, and may acquire position information using a beacon, for example.
 [外部認識センサ25]
 外部認識センサ25は、車両1の外部の状況の認識に用いられる各種のセンサを備え、各センサからのセンサデータを車両制御システム11の各部に供給する。外部認識センサ25が備えるセンサの種類や数は任意である。
[External Recognition Sensor 25]
The external recognition sensor 25 includes various sensors used to recognize the situation outside the vehicle 1, and supplies sensor data from each sensor to each unit of the vehicle control system 11. The type and number of sensors included in the external recognition sensor 25 are arbitrary.
 例えば、外部認識センサ25は、カメラ51、レーダ52、LiDAR(Light Detection and Ranging、Laser Imaging Detection and Ranging)53、及び、超音波センサ54を備える。これに限らず、外部認識センサ25は、カメラ51、レーダ52、LiDAR53、及び、超音波センサ54のうち1種類以上のセンサを備える構成でもよい。カメラ51、レーダ52、LiDAR53、及び、超音波センサ54の数は、現実的に車両1に設置可能な数であれば特に限定されない。また、外部認識センサ25が備えるセンサの種類は、この例に限定されず、外部認識センサ25は、他の種類のセンサを備えてもよい。外部認識センサ25が備える各センサのセンシング領域の例は、後述する。 For example, the external recognition sensor 25 includes a camera 51, a radar 52, a LiDAR (Light Detection and Ranging, Laser Imaging Detection and Ranging) 53, and an ultrasonic sensor 54. Without being limited to this, the external recognition sensor 25 may be configured to include one or more types of sensors among the camera 51, the radar 52, the LiDAR 53, and the ultrasonic sensor 54. The number of cameras 51, radars 52, LiDAR 53, and ultrasonic sensors 54 is not particularly limited as long as it is a number that can be realistically installed on the vehicle 1. Furthermore, the types of sensors included in the external recognition sensor 25 are not limited to this example, and the external recognition sensor 25 may include other types of sensors. Examples of the sensing areas of each sensor included in the external recognition sensor 25 will be described later.
 なお、カメラ51の撮影方式は、特に限定されない。例えば、測距が可能な撮影方式であるToF(Time Of Flight)カメラ、ステレオカメラ、単眼カメラ、赤外線カメラといった各種の撮影方式のカメラを、必要に応じてカメラ51に適用することができる。これに限らず、カメラ51は、測距に関わらずに、単に撮影画像を取得するためのものであってもよい。 The imaging method of camera 51 is not particularly limited. For example, cameras of various imaging methods, such as a ToF (Time Of Flight) camera, a stereo camera, a monocular camera, and an infrared camera, which are imaging methods capable of distance measurement, can be applied to camera 51 as necessary. However, the present invention is not limited to this, and camera 51 may simply be used to obtain a photographed image, without being related to distance measurement.
 また、例えば、外部認識センサ25は、車両1に対する環境を検出するための環境センサを備えることができる。環境センサは、天候、気象、明るさ等の環境を検出するためのセンサであって、例えば、雨滴センサ、霧センサ、日照センサ、雪センサ、照度センサ等の各種センサを含むことができる。 Furthermore, for example, the external recognition sensor 25 can be equipped with an environmental sensor for detecting the environment relative to the vehicle 1. The environmental sensor is a sensor for detecting the environment such as the weather, climate, brightness, etc., and can include various sensors such as a raindrop sensor, a fog sensor, a sunlight sensor, a snow sensor, an illuminance sensor, etc.
 さらに、例えば、外部認識センサ25は、車両1の周囲の音や音源の位置の検出等に用いられるマイクロフォンを備える。 Furthermore, for example, the external recognition sensor 25 includes a microphone that is used to detect sounds around the vehicle 1 and the location of sound sources.
 [車内センサ26]
 車内センサ26は、車内の情報を検出するための各種のセンサを備え、各センサからのセンサデータを車両制御システム11の各部に供給する。車内センサ26が備える各種センサの種類や数は、現実的に車両1に設置可能な種類や数であれば特に限定されない。
[In-vehicle sensor 26]
The in-vehicle sensor 26 includes various sensors for detecting information inside the vehicle, and supplies sensor data from each sensor to each unit of the vehicle control system 11. The types and number of the various sensors included in the in-vehicle sensor 26 are not particularly limited as long as they are of the types and number that can be realistically installed in the vehicle 1.
 例えば、車内センサ26は、カメラ、レーダ、着座センサ、ステアリングホイールセンサ、マイクロフォン、生体センサのうち1種類以上のセンサを備えることができる。車内センサ26が備えるカメラとしては、例えば、ToFカメラ、ステレオカメラ、単眼カメラ、赤外線カメラといった、測距可能な各種の撮影方式のカメラを用いることができる。これに限らず、車内センサ26が備えるカメラは、測距に関わらずに、単に撮影画像を取得するためのものであってもよい。車内センサ26が備える生体センサは、例えば、シートやステアリングホイール等に設けられ、運転者等の搭乗者の各種の生体情報を検出する。 For example, the in-vehicle sensor 26 may be equipped with one or more types of sensors including a camera, radar, a seating sensor, a steering wheel sensor, a microphone, and a biometric sensor. The camera equipped in the in-vehicle sensor 26 may be a camera using various imaging methods capable of measuring distances, such as a ToF camera, a stereo camera, a monocular camera, or an infrared camera. Without being limited to this, the camera equipped in the in-vehicle sensor 26 may be a camera simply for acquiring captured images, regardless of distance measurement. The biometric sensor equipped in the in-vehicle sensor 26 is provided, for example, on a seat, steering wheel, etc., and detects various types of biometric information of passengers such as the driver.
 [車両センサ27]
 車両センサ27は、車両1の状態を検出するための各種のセンサを備え、各センサからのセンサデータを車両制御システム11の各部に供給する。車両センサ27が備える各種センサの種類や数は、現実的に車両1に設置可能な種類や数であれば特に限定されない。
[Vehicle sensor 27]
The vehicle sensor 27 includes various sensors for detecting the state of the vehicle 1, and supplies sensor data from each sensor to each unit of the vehicle control system 11. The types and number of the various sensors included in the vehicle sensor 27 are not particularly limited as long as they are types and numbers that can be realistically installed in the vehicle 1.
 例えば、車両センサ27は、速度センサ、加速度センサ、角速度センサ(ジャイロセンサ)、及び、それらを統合した慣性計測装置(IMU(Inertial Measurement Unit))を備える。例えば、車両センサ27は、ステアリングホイールの操舵角を検出する操舵角センサ、ヨーレートセンサ、アクセルペダルの操作量を検出するアクセルセンサ、及び、ブレーキペダルの操作量を検出するブレーキセンサを備える。例えば、車両センサ27は、エンジンやモータの回転数を検出する回転センサ、タイヤの空気圧を検出する空気圧センサ、タイヤのスリップ率を検出するスリップ率センサ、及び、車輪の回転速度を検出する車輪速センサを備える。例えば、車両センサ27は、バッテリの残量及び温度を検出するバッテリセンサ、並びに、外部からの衝撃を検出する衝撃センサを備える。 For example, the vehicle sensor 27 includes a speed sensor, an acceleration sensor, an angular velocity sensor (gyro sensor), and an inertial measurement unit (IMU) that integrates these. For example, the vehicle sensor 27 includes a steering angle sensor that detects the steering angle of the steering wheel, a yaw rate sensor, an accelerator sensor that detects the amount of accelerator pedal operation, and a brake sensor that detects the amount of brake pedal operation. For example, the vehicle sensor 27 includes a rotation sensor that detects the number of rotations of the engine or motor, an air pressure sensor that detects the air pressure of the tires, a slip ratio sensor that detects the slip ratio of the tires, and a wheel speed sensor that detects the rotation speed of the wheels. For example, the vehicle sensor 27 includes a battery sensor that detects the remaining charge and temperature of the battery, and an impact sensor that detects external impacts.
 [記憶部31]
 記憶部31は、不揮発性の記憶媒体及び揮発性の記憶媒体のうち少なくとも一方を含み、データやプログラムを記憶する。記憶部31は、例えばEEPROM(Electrically Erasable Programmable Read Only Memory)及びRAM(Random Access Memory)として用いられ、記憶媒体としては、HDD(Hard Disc Drive)といった磁気記憶デバイス、半導体記憶デバイス、光記憶デバイス、及び、光磁気記憶デバイスを適用することができる。記憶部31は、車両制御システム11の各部が用いる各種プログラムやデータを記憶する。例えば、記憶部31は、EDR(Event Data Recorder)やDSSAD(Data Storage System for Automated Driving)を備え、事故等のイベントの前後の車両1の情報や車内センサ26によって取得された情報を記憶する。
[Memory unit 31]
The storage unit 31 includes at least one of a non-volatile storage medium and a volatile storage medium, and stores data and programs. The storage unit 31 is used, for example, as an electrically erasable programmable read only memory (EEPROM) and a random access memory (RAM), and the storage medium may be a magnetic storage device such as a hard disk drive (HDD), a semiconductor storage device, an optical storage device, or a magneto-optical storage device. The storage unit 31 stores various programs and data used by each part of the vehicle control system 11. For example, the storage unit 31 includes an event data recorder (EDR) and a data storage system for automated driving (DSSAD), and stores information on the vehicle 1 before and after an event such as an accident, and information acquired by the in-vehicle sensor 26.
 [走行支援・自動運転制御部32]
 走行支援・自動運転制御部32は、車両1の走行支援及び自動運転の制御を行う。例えば、走行支援・自動運転制御部32は、分析部61、行動計画部62、及び、動作制御部63を備える。
[Driving assistance/automatic driving control unit 32]
The driving assistance/automatic driving control unit 32 controls driving assistance and automatic driving of the vehicle 1. For example, the driving assistance/automatic driving control unit 32 includes an analysis unit 61, an action planning unit 62, and an operation control unit 63.
 分析部61は、車両1及び周囲の状況の分析処理を行う。分析部61は、自己位置推定部71、センサフュージョン部72、及び、認識部73を備える。 The analysis unit 61 performs analysis processing of the vehicle 1 and the surrounding conditions. The analysis unit 61 includes a self-position estimation unit 71, a sensor fusion unit 72, and a recognition unit 73.
 自己位置推定部71は、外部認識センサ25からのセンサデータ、及び、地図情報蓄積部23に蓄積されている高精度地図に基づいて、車両1の自己位置を推定する。例えば、自己位置推定部71は、外部認識センサ25からのセンサデータに基づいてローカルマップを生成し、ローカルマップと高精度地図とのマッチングを行うことにより、車両1の自己位置を推定する。車両1の位置は、例えば、後輪対車軸の中心が基準とされる。 The self-position estimation unit 71 estimates the self-position of the vehicle 1 based on the sensor data from the external recognition sensor 25 and the high-precision map stored in the map information storage unit 23. For example, the self-position estimation unit 71 generates a local map based on the sensor data from the external recognition sensor 25, and estimates the self-position of the vehicle 1 by matching the local map with the high-precision map. The position of the vehicle 1 is based on, for example, the center of the rear wheel pair axle.
 ローカルマップは、例えば、SLAM(Simultaneous Localization and Mapping)等の技術を用いて作成される3次元の高精度地図、占有格子地図(Occupancy Grid Map)等である。3次元の高精度地図は、例えば、上述したポイントクラウドマップ等である。占有格子地図は、車両1の周囲の3次元又は2次元の空間を所定の大きさのグリッド(格子)に分割し、グリッド単位で物体の占有状態を示す地図である。物体の占有状態は、例えば、物体の有無や存在確率により示される。ローカルマップは、例えば、認識部73による車両1の外部の状況の検出処理及び認識処理にも用いられる。 The local map is, for example, a three-dimensional high-precision map or an occupancy grid map created using technology such as SLAM (Simultaneous Localization and Mapping). The three-dimensional high-precision map is, for example, the point cloud map described above. The occupancy grid map is a map in which the three-dimensional or two-dimensional space around the vehicle 1 is divided into grids of a predetermined size, and the occupancy state of objects is shown on a grid-by-grid basis. The occupancy state of objects is indicated, for example, by the presence or absence of an object and the probability of its existence. The local map is also used, for example, in detection processing and recognition processing of the situation outside the vehicle 1 by the recognition unit 73.
 なお、自己位置推定部71は、位置情報取得部24により取得される位置情報、及び、車両センサ27からのセンサデータに基づいて、車両1の自己位置を推定してもよい。 The self-position estimation unit 71 may estimate the self-position of the vehicle 1 based on the position information acquired by the position information acquisition unit 24 and the sensor data from the vehicle sensor 27.
 センサフュージョン部72は、複数の異なる種類のセンサデータ(例えば、カメラ51から供給される画像データ、及び、レーダ52から供給されるセンサデータ)を組み合わせて、新たな情報を得るセンサフュージョン処理を行う。異なる種類のセンサデータを組合せる方法としては、統合、融合、連合等がある。 The sensor fusion unit 72 performs sensor fusion processing to combine multiple different types of sensor data (e.g., image data supplied from the camera 51 and sensor data supplied from the radar 52) to obtain new information. Methods for combining different types of sensor data include integration, fusion, and association.
 認識部73は、車両1の外部の状況の検出を行う検出処理、及び、車両1の外部の状況の認識を行う認識処理を実行する。 The recognition unit 73 executes a detection process to detect the situation outside the vehicle 1, and a recognition process to recognize the situation outside the vehicle 1.
 例えば、認識部73は、外部認識センサ25からの情報、自己位置推定部71からの情報、センサフュージョン部72からの情報等に基づいて、車両1の外部の状況の検出処理及び認識処理を行う。 For example, the recognition unit 73 performs detection and recognition processing of the situation outside the vehicle 1 based on information from the external recognition sensor 25, information from the self-position estimation unit 71, information from the sensor fusion unit 72, etc.
 具体的には、例えば、認識部73は、車両1の周囲の物体の検出処理及び認識処理等を行う。物体の検出処理とは、例えば、物体の有無、大きさ、形、位置、動き等を検出する処理である。物体の認識処理とは、例えば、物体の種類等の属性を認識したり、特定の物体を識別したりする処理である。ただし、検出処理と認識処理とは、必ずしも明確に分かれるものではなく、重複する場合がある。 Specifically, for example, the recognition unit 73 performs detection processing and recognition processing of objects around the vehicle 1. Object detection processing is, for example, processing to detect the presence or absence, size, shape, position, movement, etc. of an object. Object recognition processing is, for example, processing to recognize attributes such as the type of object, and to identify a specific object. However, detection processing and recognition processing are not necessarily clearly separated, and there may be overlap.
 例えば、認識部73は、レーダ52又はLiDAR53等によるセンサデータに基づくポイントクラウドを点群の塊毎に分類するクラスタリングを行うことにより、車両1の周囲の物体を検出する。これにより、車両1の周囲の物体の有無、大きさ、形状、位置が検出される。 For example, the recognition unit 73 detects objects around the vehicle 1 by performing clustering to classify a point cloud based on sensor data from the radar 52, the LiDAR 53, or the like into clusters of points. This allows the presence or absence, size, shape, and position of objects around the vehicle 1 to be detected.
 例えば、認識部73は、クラスタリングにより分類された点群の塊の動きを追従するトラッキングを行うことにより、車両1の周囲の物体の動きを検出する。これにより、車両1の周囲の物体の速度及び進行方向(移動ベクトル)が検出される。 For example, the recognition unit 73 detects the movement of objects around the vehicle 1 by performing tracking to follow the movement of clusters of point clouds classified by clustering. This allows the speed and direction of travel (movement vector) of objects around the vehicle 1 to be detected.
 例えば、認識部73は、カメラ51から供給される画像データに基づいて、車両、人、自転車、障害物、構造物、道路、信号機、交通標識、道路標示等を検出又は認識する。また、認識部73は、セマンティックセグメンテーション等の認識処理を行うことにより、車両1の周囲の物体の種類を認識してもよい。 For example, the recognition unit 73 detects or recognizes vehicles, people, bicycles, obstacles, structures, roads, traffic lights, traffic signs, road markings, etc. based on image data supplied from the camera 51. The recognition unit 73 may also recognize the types of objects around the vehicle 1 by performing recognition processing such as semantic segmentation.
 例えば、認識部73は、地図情報蓄積部23に蓄積されている地図、自己位置推定部71による自己位置の推定結果、及び、認識部73による車両1の周囲の物体の認識結果に基づいて、車両1の周囲の交通ルールの認識処理を行うことができる。認識部73は、この処理により、信号機の位置及び状態、交通標識及び道路標示の内容、交通規制の内容、並びに、走行可能な車線等を認識することができる。 For example, the recognition unit 73 can perform recognition processing of traffic rules around the vehicle 1 based on the map stored in the map information storage unit 23, the result of self-location estimation by the self-location estimation unit 71, and the result of recognition of objects around the vehicle 1 by the recognition unit 73. Through this processing, the recognition unit 73 can recognize the positions and states of traffic lights, the contents of traffic signs and road markings, the contents of traffic regulations, and lanes on which travel is possible, etc.
 例えば、認識部73は、車両1の周囲の環境の認識処理を行うことができる。認識部73が認識対象とする周囲の環境としては、天候、気温、湿度、明るさ、及び、路面の状態等が想定される。 For example, the recognition unit 73 can perform recognition processing of the environment around the vehicle 1. The surrounding environment that the recognition unit 73 recognizes may include weather, temperature, humidity, brightness, and road surface conditions.
 行動計画部62は、車両1の行動計画を作成する。例えば、行動計画部62は、経路計画、経路追従の処理を行うことにより、行動計画を作成する。 The behavior planning unit 62 creates a behavior plan for the vehicle 1. For example, the behavior planning unit 62 creates the behavior plan by performing route planning and route following processing.
 なお、経路計画(Global path planning)とは、スタートからゴールまでの大まかな経路を計画する処理である。この経路計画には、軌道計画と言われ、計画した経路において、車両1の運動特性を考慮して、車両1の近傍で安全かつ滑らかに進行することが可能な軌道生成(Local path planning)を行う処理も含まれる。 Global path planning is a process that plans a rough route from the start to the goal. This route planning is called trajectory planning, and also includes a process of local path planning that takes into account the motion characteristics of vehicle 1 on the planned route and generates a trajectory that allows safe and smooth progress in the vicinity of vehicle 1.
 経路追従とは、経路計画により計画された経路を計画された時間内で安全かつ正確に走行するための動作を計画する処理である。行動計画部62は、例えば、この経路追従の処理の結果に基づき、車両1の目標速度と目標角速度を計算することができる。 Path following is a process of planning operations for traveling safely and accurately along a route planned by a route plan within a planned time. The action planning unit 62 can, for example, calculate the target speed and target angular velocity of the vehicle 1 based on the results of this path following process.
 動作制御部63は、行動計画部62により作成された行動計画を実現するために、車両1の動作を制御する。 The operation control unit 63 controls the operation of the vehicle 1 to realize the action plan created by the action planning unit 62.
 例えば、動作制御部63は、後述する車両制御部35に含まれる、ステアリング制御部81、ブレーキ制御部82、及び、駆動制御部83を制御して、軌道計画により計算された軌道を車両1が進行するように、加減速制御及び方向制御を行う。例えば、動作制御部63は、衝突回避又は衝撃緩和、追従走行、車速維持走行、自車の衝突警告、自車のレーン逸脱警告等のADASの機能実現を目的とした協調制御を行う。例えば、動作制御部63は、運転者の操作によらずに自律的に走行する自動運転等を目的とした協調制御を行う。 For example, the operation control unit 63 controls the steering control unit 81, the brake control unit 82, and the drive control unit 83 included in the vehicle control unit 35 described below, and performs acceleration/deceleration control and directional control so that the vehicle 1 proceeds along the trajectory calculated by the trajectory plan. For example, the operation control unit 63 performs cooperative control aimed at realizing ADAS functions such as collision avoidance or impact mitigation, following driving, maintaining vehicle speed, collision warning for the vehicle itself, and lane departure warning for the vehicle itself. For example, the operation control unit 63 performs cooperative control aimed at automatic driving, which drives autonomously without the driver's operation.
 [DMS33]
 DMS33は、車内センサ26からのセンサデータ、及び、後述するHMI34に入力される入力データ等に基づいて、運転者の認証処理、及び、運転者の状態の認識処理等を行う。認識対象となる運転者の状態としては、例えば、体調、覚醒度、集中度、疲労度、視線方向、酩酊度、運転操作、姿勢等が想定される。
[DMS33]
The DMS 33 performs authentication processing of the driver and recognition processing of the driver's state based on the sensor data from the in-vehicle sensor 26 and input data input to the HMI 34 (described later), etc. Examples of the driver's state to be recognized include physical condition, alertness level, concentration level, fatigue level, line of sight direction, level of intoxication, driving operation, posture, etc.
 なお、DMS33が、運転者以外の搭乗者の認証処理、及び、当該搭乗者の状態の認識処理を行うようにしてもよい。また、例えば、DMS33が、車内センサ26からのセンサデータに基づいて、車内の状況の認識処理を行うようにしてもよい。認識対象となる車内の状況としては、例えば、気温、湿度、明るさ、臭い等が想定される。 The DMS 33 may also perform authentication processing for passengers other than the driver and recognition processing for the status of the passengers. For example, the DMS 33 may also perform recognition processing for the situation inside the vehicle based on sensor data from the in-vehicle sensor 26. Examples of the situation inside the vehicle that may be recognized include temperature, humidity, brightness, odor, etc.
 [HMI34]
 HMI34は、各種のデータや指示等の入力と、各種のデータの運転者等への提示を行う。
[HMI34]
The HMI 34 inputs various data and instructions, and presents various data to the driver, etc.
 HMI34によるデータの入力について、概略的に説明する。HMI34は、人がデータを入力するための入力デバイスを備える。HMI34は、入力デバイスにより入力されたデータや指示等に基づいて入力信号を生成し、車両制御システム11の各部に供給する。HMI34は、入力デバイスとして、例えばタッチパネル、ボタン、スイッチ、及び、レバーといった操作子を備える。これに限らず、HMI34は、音声やジェスチャ等により手動操作以外の方法で情報を入力可能な入力デバイスをさらに備えてもよい。さらに、HMI34は、例えば、赤外線又は電波を利用したリモートコントロール装置や、車両制御システム11の操作に対応したモバイル機器又はウェアラブル機器等の外部接続機器を入力デバイスとして用いてもよい。 The following provides an overview of data input by the HMI 34. The HMI 34 is equipped with an input device that allows a person to input data. The HMI 34 generates input signals based on data and instructions input by the input device, and supplies the signals to each part of the vehicle control system 11. The HMI 34 is equipped with input devices such as a touch panel, buttons, switches, and levers. Without being limited to these, the HMI 34 may further be equipped with an input device that allows information to be input by a method other than manual operation, such as voice or gestures. Furthermore, the HMI 34 may use, as an input device, an externally connected device such as a remote control device that uses infrared rays or radio waves, or a mobile device or wearable device that supports the operation of the vehicle control system 11.
 HMI34によるデータの提示について、概略的に説明する。HMI34は、搭乗者又は車外に対する視覚情報、聴覚情報、及び、触覚情報の生成を行う。また、HMI34は、生成された各情報の出力、出力内容、出力タイミング及び出力方法等を制御する出力制御を行う。HMI34は、視覚情報として、例えば、操作画面、車両1の状態表示、警告表示、車両1の周囲の状況を示すモニタ画像等の画像や光により示される情報を生成及び出力する。また、HMI34は、聴覚情報として、例えば、音声ガイダンス、警告音、警告メッセージ等の音により示される情報を生成及び出力する。さらに、HMI34は、触覚情報として、例えば、力、振動、動き等により搭乗者の触覚に与えられる情報を生成及び出力する。 The presentation of data by the HMI 34 will be briefly described below. The HMI 34 generates visual information, auditory information, and tactile information for the occupant or the outside of the vehicle. The HMI 34 also performs output control to control the output, output content, output timing, output method, etc. of each piece of generated information. The HMI 34 generates and outputs, as visual information, information indicated by images or light, such as an operation screen, a status display of the vehicle 1, a warning display, and a monitor image showing the situation around the vehicle 1. The HMI 34 also generates and outputs, as auditory information, information indicated by sounds, such as voice guidance, warning sounds, and warning messages. The HMI 34 also generates and outputs, as tactile information, information that is imparted to the occupant's sense of touch by, for example, force, vibration, movement, etc.
 HMI34が視覚情報を出力する出力デバイスとしては、例えば、自身が画像を表示することで視覚情報を提示する表示装置や、画像を投影することで視覚情報を提示するプロジェクタ装置を適用することができる。なお、表示装置は、通常のディスプレイを有する表示装置以外にも、例えば、ヘッドアップディスプレイ、透過型ディスプレイ、AR(Augmented Reality)機能を備えるウエアラブルデバイスといった、搭乗者の視界内に視覚情報を表示する装置であってもよい。また、HMI34は、車両1に設けられるナビゲーション装置、インストルメントパネル、CMS(Camera Monitoring System)、電子ミラー、ランプ等が有する表示デバイスを、視覚情報を出力する出力デバイスとして用いることも可能である。 The output device from which the HMI 34 outputs visual information may be, for example, a display device that presents visual information by displaying an image itself, or a projector device that presents visual information by projecting an image. Note that the display device may be a device that displays visual information within the field of vision of the passenger, such as a head-up display, a transmissive display, or a wearable device with an AR (Augmented Reality) function, in addition to a display device having a normal display. The HMI 34 may also use display devices such as a navigation device, an instrument panel, a CMS (Camera Monitoring System), an electronic mirror, or lamps provided in the vehicle 1 as output devices that output visual information.
 HMI34が聴覚情報を出力する出力デバイスとしては、例えば、オーディオスピーカ、ヘッドホン、イヤホンを適用することができる。 The output device through which the HMI 34 outputs auditory information can be, for example, an audio speaker, headphones, or earphones.
 HMI34が触覚情報を出力する出力デバイスとしては、例えば、ハプティクス技術を用いたハプティクス素子を適用することができる。ハプティクス素子は、例えば、ステアリングホイール、シートといった、車両1の搭乗者が接触する部分に設けられる。 Haptic elements using haptic technology can be used as output devices for the HMI 34 to output haptic information. Haptic elements are provided on parts of the vehicle 1 that are in contact with passengers, such as the steering wheel and the seat.
 [車両制御部35]
 車両制御部35は、車両1の各部の制御を行う。車両制御部35は、ステアリング制御部81、ブレーキ制御部82、駆動制御部83、ボディ系制御部84、ライト制御部85、及び、ホーン制御部86を備える。
[Vehicle control unit 35]
The vehicle control unit 35 controls each unit of the vehicle 1. The vehicle control unit 35 includes a steering control unit 81, a brake control unit 82, a drive control unit 83, a body system control unit 84, a light control unit 85, and a horn control unit 86.
 ステアリング制御部81は、車両1のステアリングシステムの状態の検出及び制御等を行う。ステアリングシステムは、例えば、ステアリングホイール等を備えるステアリング機構、電動パワーステアリング等を備える。ステアリング制御部81は、例えば、ステアリングシステムの制御を行うステアリングECU、ステアリングシステムの駆動を行うアクチュエータ等を備える。 The steering control unit 81 detects and controls the state of the steering system of the vehicle 1. The steering system includes, for example, a steering mechanism including a steering wheel, an electric power steering, etc. The steering control unit 81 includes, for example, a steering ECU that controls the steering system, an actuator that drives the steering system, etc.
 ブレーキ制御部82は、車両1のブレーキシステムの状態の検出及び制御等を行う。ブレーキシステムは、例えば、ブレーキペダル等を含むブレーキ機構、ABS(Antilock Brake System)、回生ブレーキ機構等を備える。ブレーキ制御部82は、例えば、ブレーキシステムの制御を行うブレーキECU、ブレーキシステムの駆動を行うアクチュエータ等を備える。 The brake control unit 82 detects and controls the state of the brake system of the vehicle 1. The brake system includes, for example, a brake mechanism including a brake pedal, an ABS (Antilock Brake System), a regenerative brake mechanism, etc. The brake control unit 82 includes, for example, a brake ECU that controls the brake system, and an actuator that drives the brake system.
 駆動制御部83は、車両1の駆動システムの状態の検出及び制御等を行う。駆動システムは、例えば、アクセルペダル、内燃機関又は駆動用モータ等の駆動力を発生させるための駆動力発生装置、駆動力を車輪に伝達するための駆動力伝達機構等を備える。駆動制御部83は、例えば、駆動システムの制御を行う駆動ECU、駆動システムの駆動を行うアクチュエータ等を備える。 The drive control unit 83 detects and controls the state of the drive system of the vehicle 1. The drive system includes, for example, an accelerator pedal, a drive force generating device for generating drive force such as an internal combustion engine or a drive motor, and a drive force transmission mechanism for transmitting the drive force to the wheels. The drive control unit 83 includes, for example, a drive ECU for controlling the drive system, and an actuator for driving the drive system.
 ボディ系制御部84は、車両1のボディ系システムの状態の検出及び制御等を行う。ボディ系システムは、例えば、キーレスエントリシステム、スマートキーシステム、パワーウインドウ装置、パワーシート、空調装置、エアバッグ、シートベルト、シフトレバー等を備える。ボディ系制御部84は、例えば、ボディ系システムの制御を行うボディ系ECU、ボディ系システムの駆動を行うアクチュエータ等を備える。 The body system control unit 84 detects and controls the state of the body system of the vehicle 1. The body system includes, for example, a keyless entry system, a smart key system, a power window device, a power seat, an air conditioning system, an airbag, a seat belt, a shift lever, etc. The body system control unit 84 includes, for example, a body system ECU that controls the body system, an actuator that drives the body system, etc.
 ライト制御部85は、車両1の各種のライトの状態の検出及び制御等を行う。制御対象となるライトとしては、例えば、ヘッドライト、バックライト、フォグライト、ターンシグナル、ブレーキライト、プロジェクション、バンパーの表示等が想定される。ライト制御部85は、ライトの制御を行うライトECU、ライトの駆動を行うアクチュエータ等を備える。 The light control unit 85 detects and controls the state of various lights of the vehicle 1. Examples of lights to be controlled include headlights, backlights, fog lights, turn signals, brake lights, projection, and bumper displays. The light control unit 85 includes a light ECU that controls the lights, an actuator that drives the lights, and the like.
 ホーン制御部86は、車両1のカーホーンの状態の検出及び制御等を行う。ホーン制御部86は、例えば、カーホーンの制御を行うホーンECU、カーホーンの駆動を行うアクチュエータ等を備える。 The horn control unit 86 detects and controls the state of the car horn of the vehicle 1. The horn control unit 86 includes, for example, a horn ECU that controls the car horn, an actuator that drives the car horn, etc.
 図2は、第1実施形態の車両1のセンシング領域を示す平面図である。図2は、図1の外部認識センサ25のカメラ51、レーダ52、LiDAR53、及び、超音波センサ54等によるセンシング領域の例を示している。なお、図2において、車両1を上面から見た様子が模式的に示され、左端側が車両1の前端(フロント)側であり、右端側が車両1の後端(リア)側となっている。 FIG. 2 is a plan view showing the sensing area of the vehicle 1 in the first embodiment. FIG. 2 shows an example of the sensing area of the camera 51, radar 52, LiDAR 53, ultrasonic sensor 54, etc. of the external recognition sensor 25 in FIG. 1. Note that FIG. 2 shows a schematic view of the vehicle 1 as seen from above, with the left end side being the front end of the vehicle 1 and the right end side being the rear end of the vehicle 1.
 [センシング領域1-1F,B]
 センシング領域1-1F及びセンシング領域1-1Bは、超音波センサ54のセンシング領域の例を示している。センシング領域1-1Fは、複数の超音波センサ54によって車両1の前端周辺をカバーしている。センシング領域1-1Bは、複数の超音波センサ54によって車両1の後端周辺をカバーしている。
[Sensing area 1-1F, B]
Sensing area 1-1F and sensing area 1-1B are examples of sensing areas of the ultrasonic sensors 54. The sensing area 1-1F covers the periphery of the front end of the vehicle 1 with a plurality of ultrasonic sensors 54. The sensing area 1-1B covers the periphery of the rear end of the vehicle 1 with a plurality of ultrasonic sensors 54.
 センシング領域1-1F及びセンシング領域1-1Bにおけるセンシング結果は、例えば、車両1の駐車支援等に用いられる。 The sensing results in sensing area 1-1F and sensing area 1-1B are used, for example, for parking assistance for vehicle 1.
 [センシング領域1-2F,B,L,R]
 センシング領域1-2F乃至センシング領域1-2Bは、短距離又は中距離用のレーダ52のセンシング領域の例を示している。センシング領域1-2Fは、車両1の前方において、センシング領域1-1Fより遠い位置までカバーしている。センシング領域1-2Bは、車両1の後方において、センシング領域1-1Bより遠い位置までカバーしている。センシング領域1-2Lは、車両1の左側面の後方の周辺をカバーしている。センシング領域1-2Rは、車両1の右側面の後方の周辺をカバーしている。
[Sensing areas 1-2F, B, L, R]
Sensing area 1-2F to sensing area 1-2B show examples of sensing areas of a short-range or medium-range radar 52. Sensing area 1-2F covers a position farther in front of the vehicle 1 than sensing area 1-1F. Sensing area 1-2B covers a position farther in the rear of the vehicle 1 than sensing area 1-1B. Sensing area 1-2L covers the rear periphery of the left side of the vehicle 1. Sensing area 1-2R covers the rear periphery of the right side of the vehicle 1.
 センシング領域1-2Fにおけるセンシング結果は、例えば、車両1の前方に存在する車両や歩行者等の検出等に用いられる。センシング領域1-2Bにおけるセンシング結果は、例えば、車両1の後方の衝突防止機能等に用いられる。センシング領域1-2L及びセンシング領域1-2Rにおけるセンシング結果は、例えば、車両1の側方の死角における物体の検出等に用いられる。 The sensing results in sensing area 1-2F are used, for example, to detect vehicles, pedestrians, etc. in front of vehicle 1. The sensing results in sensing area 1-2B are used, for example, for collision prevention functions behind vehicle 1. The sensing results in sensing area 1-2L and sensing area 1-2R are used, for example, to detect objects in blind spots to the sides of vehicle 1.
 [センシング領域1-3F,B,L,R]
 センシング領域1-3F乃至センシング領域1-3Bは、カメラ51によるセンシング領域の例を示している。センシング領域1-3Fは、車両1の前方において、センシング領域1-2Fより遠い位置までカバーしている。センシング領域1-3Bは、車両1の後方において、センシング領域1-2Bより遠い位置までカバーしている。センシング領域1-3Lは、車両1の左側面の周辺をカバーしている。センシング領域1-3Rは、車両1の右側面の周辺をカバーしている。
[Sensing areas 1-3F, B, L, R]
Sensing area 1-3F to sensing area 1-3B show examples of sensing areas sensed by camera 51. Sensing area 1-3F covers a position farther in front of vehicle 1 than sensing area 1-2F. Sensing area 1-3B covers a position farther in the rear of vehicle 1 than sensing area 1-2B. Sensing area 1-3L covers the periphery of the left side of vehicle 1. Sensing area 1-3R covers the periphery of the right side of vehicle 1.
 センシング領域1-3Fにおけるセンシング結果は、例えば、信号機や交通標識の認識、車線逸脱防止支援システム、自動ヘッドライト制御システムに用いることができる。センシング領域1-3Bにおけるセンシング結果は、例えば、駐車支援、及び、サラウンドビューシステムに用いることができる。センシング領域1-3L及びセンシング領域1-3Rにおけるセンシング結果は、例えば、サラウンドビューシステムに用いることができる。 The sensing results in sensing area 1-3F can be used, for example, for recognizing traffic lights and traffic signs, lane departure prevention support systems, and automatic headlight control systems. The sensing results in sensing area 1-3B can be used, for example, for parking assistance and surround view systems. The sensing results in sensing area 1-3L and sensing area 1-3R can be used, for example, for surround view systems.
 [センシング領域1-4]
 センシング領域1-4は、LiDAR53のセンシング領域の例を示している。センシング領域1-4は、車両1の前方において、センシング領域1-3Fより遠い位置までカバーしている。一方、センシング領域1-4は、センシング領域1-3Fより左右方向の範囲が狭くなっている。
[Sensing area 1-4]
A sensing area 1-4 shows an example of a sensing area of the LiDAR 53. The sensing area 1-4 covers a position farther in front of the vehicle 1 than the sensing area 1-3F. On the other hand, the sensing area 1-4 has a narrower range in the left-right direction than the sensing area 1-3F.
 センシング領域1-4におけるセンシング結果は、例えば、周辺車両等の物体検出に用いられる。 The sensing results in sensing areas 1-4 are used, for example, to detect objects such as nearby vehicles.
 [センシング領域1-5]
 センシング領域1-5は、長距離用のレーダ52のセンシング領域の例を示している。センシング領域1-5は、車両1の前方において、センシング領域1-4より遠い位置までカバーしている。一方、センシング領域1-5は、センシング領域1-4より左右方向の範囲が狭くなっている。
[Sensing Areas 1-5]
A sensing area 1-5 shows an example of a sensing area of a long-range radar 52. The sensing area 1-5 covers a position farther ahead of the vehicle 1 than the sensing area 1-4. On the other hand, the sensing area 1-5 has a narrower range in the left-right direction than the sensing area 1-4.
 センシング領域1-5におけるセンシング結果は、例えば、ACC(Adaptive Cruise Control)、緊急ブレーキ、衝突回避等に用いられる。 The sensing results in sensing areas 1-5 are used, for example, for ACC (Adaptive Cruise Control), emergency braking, collision avoidance, etc.
 なお、外部認識センサ25が含むカメラ51、レーダ52、LiDAR53、及び、超音波センサ54の各センサのセンシング領域は、図2以外に各種の構成をとってもよい。具体的には、超音波センサ54が車両1の側方もセンシングするようにしてもよいし、LiDAR53が車両1の後方をセンシングするようにしてもよい。また、各センサの設置位置は、上述した各例に限定されない。また、各センサの数は、1つでもよいし、複数であってもよい。 The sensing areas of the cameras 51, radar 52, LiDAR 53, and ultrasonic sensors 54 included in the external recognition sensor 25 may have various configurations other than those shown in FIG. 2. Specifically, the ultrasonic sensor 54 may also sense the sides of the vehicle 1, and the LiDAR 53 may sense the rear of the vehicle 1. The installation positions of the sensors are not limited to the examples described above. The number of sensors may be one or more.
 (2)第1実施形態の固体撮像装置100
 図3は、第1実施形態の固体撮像装置100の構成を示すブロック図である。
(2) Solid-state imaging device 100 according to the first embodiment
FIG. 3 is a block diagram showing the configuration of the solid-state imaging device 100 according to the first embodiment.
 固体撮像装置100は、図1に示す車両1に設けられており、例えば、外部認識センサ25に含まれている。固体撮像装置100は、被写体の変化を検出するためのEVSである。被写体の例は、車両1の前方に存在する人間、車両、障害物などである。なお、固体撮像装置100は、後述する例のように、スマートフォンなどの電子機器200に内蔵されていてもよいし、ゲームコンソールなどの電子機器200に電気的に接続されていてもよい(図10および図11を参照)。 The solid-state imaging device 100 is provided in the vehicle 1 shown in FIG. 1, and is included in the external recognition sensor 25, for example. The solid-state imaging device 100 is an EVS for detecting changes in a subject. Examples of subjects include a person, a vehicle, and an obstacle in front of the vehicle 1. The solid-state imaging device 100 may be built into an electronic device 200 such as a smartphone, as in an example described later, or may be electrically connected to an electronic device 200 such as a game console (see FIGS. 10 and 11).
 固体撮像装置100は、図3に示すように、画素アレイ101と、イベント取得部102と、イベント生成部103と、イベント合成部104と、イベント出力部105とを備えている。画素アレイ101は、複数の画素101aを含んでいる。イベント生成部103は、第1フィルタ部103aと、第2フィルタ部103bと、第3フィルタ部103cとを含んでいる。イベント合成部104は、オクターブ情報付加部104aと、出力タイミング調整部104bとを含んでいる。イベント出力部105は、イベントデータ選択部105aと、イベントデータ形成部105bとを含んでいる。 As shown in FIG. 3, the solid-state imaging device 100 includes a pixel array 101, an event acquisition unit 102, an event generation unit 103, an event synthesis unit 104, and an event output unit 105. The pixel array 101 includes a plurality of pixels 101a. The event generation unit 103 includes a first filter unit 103a, a second filter unit 103b, and a third filter unit 103c. The event synthesis unit 104 includes an octave information addition unit 104a and an output timing adjustment unit 104b. The event output unit 105 includes an event data selection unit 105a and an event data formation unit 105b.
 [画素アレイ101]
 画素アレイ101は、2次元アレイ状(マトリクス状)に配置された複数の画素101aを含んでいる。図3において、紙面上の横方向(水平方向)は、画素アレイ101の行方向に対応しており、紙面上の縦方向(垂直方向)は、画素アレイ101の列方向に対応している。
[Pixel array 101]
The pixel array 101 includes a plurality of pixels 101a arranged in a two-dimensional array (matrix). In FIG. 3, the horizontal direction on the paper corresponds to the row direction of the pixel array 101, and the vertical direction on the paper corresponds to the column direction of the pixel array 101.
 各画素101aは、オンイベントやオフイベントなどのイベントを検出する機能を有する。オンイベントは、画素101aの輝度が増加した場合において、輝度の変化量(増加量)の絶対値が閾値よりも大きい場合に発火する。オフイベントは、画素101aの輝度が減少した場合において、輝度の変化量(減少量)の絶対値が閾値よりも大きい場合に発火する。例えば、オンイベントは、画素101aに被写体が入ってきた際に発火し、オフイベントは、画素101aから被写体が出ていった際に発火する。そして、各画素101aは、イベントの検出結果を示すイベントデータを出力する。 Each pixel 101a has the function of detecting events such as on events and off events. An on event is fired when the luminance of the pixel 101a increases and the absolute value of the amount of change (increase) in luminance is greater than a threshold value. An off event is fired when the luminance of the pixel 101a decreases and the absolute value of the amount of change (decrease) in luminance is greater than a threshold value. For example, an on event is fired when a subject enters the pixel 101a, and an off event is fired when a subject leaves the pixel 101a. Each pixel 101a then outputs event data indicating the event detection result.
 [イベント取得部102]
 イベント取得部102は、画素アレイ101内の各画素101aからイベント(イベントデータ)を取得する。本実施形態では、固体撮像装置100がスキャン型EVSであるため、イベント取得部102は、画素アレイ101の複数の行を順番にスキャンし、画素アレイ101からイベントデータを行ごとに取得する。
[Event Acquisition Unit 102]
The event acquiring unit 102 acquires an event (event data) from each pixel 101a in the pixel array 101. In this embodiment, since the solid-state imaging device 100 is a scan-type EVS, the event acquiring unit 102 sequentially scans multiple rows of the pixel array 101 and acquires event data from the pixel array 101 for each row.
 画素アレイ101の各行のイベントデータは、第1オクターブのイベントデータV1として取り扱われる。イベント取得部102は、第1オクターブのイベントデータV1を、イベント生成部103およびイベント合成部104に出力する。本実施形態では、イベント取得部102が取得した第1オクターブのイベントデータV1が、イベント合成部104に送られ、その複製が、イベント生成部103に送られる。 The event data for each row of the pixel array 101 is treated as event data V1 of the first octave. The event acquisition unit 102 outputs the event data V1 of the first octave to the event generation unit 103 and the event synthesis unit 104. In this embodiment, the event data V1 of the first octave acquired by the event acquisition unit 102 is sent to the event synthesis unit 104, and a copy of the event data V1 is sent to the event generation unit 103.
 図3は、第1オクターブのイベントデータV1を、一列に並んだ複数の領域P1で模式的に示している。各領域P1は、第1オクターブのイベントデータV1における、1画素101a分のイベントデータを示している。例えば、4つの領域P1は、4画素101a分のイベントデータに対応している。第1オクターブのイベントデータV1を、第1オクターブのイベント発火有無列とも呼ぶ。第1オクターブのイベントデータV1は、1行分の画素101aのイベントデータを表す文字列のデータとなっている。本実施形態では、ある画素101aでイベントが発火したか否かを、その画素101aのイベントデータを取得することで知ることができる。 FIG. 3 shows the first octave event data V1 as a series of multiple regions P1 arranged in a row. Each region P1 represents the event data for one pixel 101a in the first octave event data V1. For example, four regions P1 correspond to event data for four pixels 101a. The first octave event data V1 is also called the first octave event firing/non-firing column. The first octave event data V1 is character string data representing the event data for one row of pixels 101a. In this embodiment, it is possible to know whether an event has fired at a pixel 101a by acquiring the event data for that pixel 101a.
 [イベント生成部103]
 イベント生成部103は、イベント取得部102から出力されたイベントデータを第1オクターブのイベントデータV1とし、第1オクターブのイベントデータV1から第2~第iオクターブ(iは2以上の整数)のイベントデータV2~Viを生成する。図3は、i=4の例、すなわち、第1オクターブのイベントデータV1から第2~第4オクターブのイベントデータV2~V4を生成する例を示している。
[Event generating unit 103]
The event generating unit 103 regards the event data output from the event acquiring unit 102 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves (i is an integer equal to or greater than 2) from the event data V1 of the first octave. Fig. 3 shows an example where i = 4, that is, an example where event data V2 to V4 of the second to fourth octaves are generated from the event data V1 of the first octave.
 第1フィルタ部103aは、第1オクターブのイベントデータV1を受信すると、m行分(mは2以上の整数)の第1オクターブのイベントデータV1が溜まるまで待機する。m行分の第1オクターブのイベントデータV1は、画素アレイ101におけるm行分の画素101aのイベントデータに対応している。第1フィルタ部103aは、m行分の第1オクターブのイベントデータV1が溜まると、m行n列分(nは2以上の整数)の第1オクターブのイベントデータV1から、r行s列分(rはr<mを満たす整数、sはs<nを満たす整数)の第2オクターブのイベントデータV2を生成する。第1フィルタ部103aはその後、上記のm行n列分の第1オクターブのイベントデータV1を破棄する。第1フィルタ部103aは、このような処理をすべての行の第1オクターブのイベントデータV1について順番に繰り返す。 When the first filter unit 103a receives the first octave event data V1, it waits until m1 rows ( m1 is an integer equal to or greater than 2) of the first octave event data V1 are accumulated. The m1 rows of the first octave event data V1 correspond to the event data of m1 rows of the pixels 101a in the pixel array 101. When the m1 rows of the first octave event data V1 are accumulated, the first filter unit 103a generates r1 rows and s1 columns of the second octave event data V2 ( r1 is an integer that satisfies r1 < m1 , s1 is an integer that satisfies s1 < n1 ) from the m1 rows and n1 columns of the first octave event data V1 ( n1 is an integer equal to or greater than 2). The first filter unit 103a then discards the m1 rows and n1 columns of the first octave event data V1. The first filter unit 103a repeats this process for the event data V1 of the first octave of all rows in order.
 第1フィルタ部103aは、r行s列分の第2オクターブのイベントデータV2を、第2フィルタ部103bおよびイベント合成部104に出力する。本実施形態では、第1フィルタ部103aが生成した第2オクターブのイベントデータV2が、イベント合成部104に送られ、その複製が、第2フィルタ部103bに送られる。 The first filter unit 103a outputs the second octave event data V2 of r 1 rows and s 1 columns to the second filter unit 103b and the event synthesis unit 104. In this embodiment, the second octave event data V2 generated by the first filter unit 103a is sent to the event synthesis unit 104, and a copy thereof is sent to the second filter unit 103b.
 図3は、第2オクターブのイベントデータV2を、一列に並んだ複数の領域P2で模式的に示している。図3は、m=2、n=2、r=1、s=1の例を示している。そのため、2行2列分の領域P1が1行1列分の領域P2に置換されている。ここで、本実施形態の画素アレイ101は、M行N列分の画素101aを含んでいるものとする。そのため、第1オクターブのイベントデータV1から第2オクターブのイベントデータV2を生成する処理は、全部でM×N/4回繰り返される。図3では、4つの領域P1が1つの領域P2に集約されており、従って、各領域P2が、4画素101a分のイベントデータを集約したイベントデータに対応している。第2オクターブのイベントデータV2を、第2オクターブのイベント発火有無列とも呼ぶ。 FIG. 3 shows the second octave event data V2 as a schematic representation of a plurality of regions P2 arranged in a row. FIG. 3 shows an example where m 1 =2, n 1 =2, r 1 =1, and s 1 =1. Therefore, the region P1 of 2 rows and 2 columns is replaced with the region P2 of 1 row and 1 column. Here, it is assumed that the pixel array 101 of this embodiment includes pixels 101a of M rows and N columns. Therefore, the process of generating the event data V2 of the second octave from the event data V1 of the first octave is repeated M×N/4 times in total. In FIG. 3, four regions P1 are aggregated into one region P2, and therefore each region P2 corresponds to event data obtained by aggregating the event data of four pixels 101a. The event data V2 of the second octave is also called the event ignition presence/absence column of the second octave.
 第2フィルタ部103bは、第2オクターブのイベントデータV2を受信すると、m行分(mは2以上の整数)の第2オクターブのイベントデータV2が溜まるまで待機する。第2フィルタ部103bは、m行分の第1オクターブのイベントデータV2が溜まると、m行n列分(nは2以上の整数)の第2オクターブのイベントデータV2から、r行s列分(rはr<mを満たす整数、sはs<nを満たす整数)の第3オクターブのイベントデータV3を生成する。第2フィルタ部103bはその後、上記のm行n列分の第1オクターブのイベントデータV2を破棄する。第2フィルタ部103bは、このような処理をすべての行の第2オクターブのイベントデータV2について順番に繰り返す。 When the second filter unit 103b receives the second octave event data V2, it waits until m 2 rows (m 2 is an integer equal to or greater than 2) of the second octave event data V2 are accumulated. When m 2 rows of the first octave event data V2 are accumulated, the second filter unit 103b generates r 2 rows and s 2 columns (r 2 is an integer that satisfies r 2 <m 2 , and s 2 is an integer that satisfies s 2 <n 2 ) of the third octave event data V3 from m 2 rows and n 2 columns (n 2 is an integer equal to or greater than 2 ) of the second octave event data V2. The second filter unit 103b then discards the m 2 rows and n 2 columns of the first octave event data V2. The second filter unit 103b repeats such processing for all rows of the second octave event data V2 in order.
 第2フィルタ部103bは、r行s列分の第3オクターブのイベントデータV3を、第3フィルタ部103cおよびイベント合成部104に出力する。本実施形態では、第2フィルタ部103bが生成した第2オクターブのイベントデータV3が、イベント合成部104に送られ、その複製が、第3フィルタ部103cに送られる。 The second filter unit 103b outputs the third octave event data V3 of r 2 rows and s 2 columns to the third filter unit 103c and the event synthesis unit 104. In this embodiment, the second octave event data V3 generated by the second filter unit 103b is sent to the event synthesis unit 104, and a copy thereof is sent to the third filter unit 103c.
 図3は、第3オクターブのイベントデータV3を、一列に並んだ複数の領域P3で模式的に示している。図3は、m=2、n=2、r=1、s=1の例を示している。そのため、2行2列分の領域P2が1行1列分の領域P3に置換されている。第2オクターブのイベントデータV2から第3オクターブのイベントデータV3を生成する処理は、全部でM×N/16回繰り返される。図3では、4つの領域P2が1つの領域P3に集約されており、従って、各領域P3が、16画素101a分のイベントデータを集約したイベントデータに対応している。第3オクターブのイベントデータV3を、第3オクターブのイベント発火有無列とも呼ぶ。 FIG. 3 shows the third octave event data V3 as a schematic representation of a plurality of regions P3 arranged in a row. FIG. 3 shows an example where m 2 =2, n 2 =2, r 2 =1, and s 2 =1. Therefore, the region P2 of 2 rows and 2 columns is replaced with the region P3 of 1 row and 1 column. The process of generating the third octave event data V3 from the second octave event data V2 is repeated M×N/16 times in total. In FIG. 3, four regions P2 are aggregated into one region P3, and therefore each region P3 corresponds to event data obtained by aggregating event data for 16 pixels 101a. The third octave event data V3 is also called the third octave event ignition presence/absence column.
 第3フィルタ部103cの動作は、第1フィルタ部103aや第2フィルタ部103bの動作と同様である。第3フィルタ部103cは、m行n列分(nは2以上の整数、nは2以上の整数)の第3オクターブのイベントデータV3から、r行s列分(rはr<mを満たす整数、sはs<nを満たす整数)の第4オクターブのイベントデータV4を生成する。第3フィルタ部103cは、r行s列分の第4オクターブのイベントデータV4を、イベント合成部104に出力する。図3は、第4オクターブのイベントデータV4を、一列に並んだ複数の領域P4で模式的に示している。図3は、m=2、n=2、r=1、s=1の例を示している。そのため、2行2列分の領域P3が1行1列分の領域P4に置換されている。第4オクターブのイベントデータV4を、第4オクターブのイベント発火有無列とも呼ぶ。 The operation of the third filter unit 103c is the same as that of the first filter unit 103a and the second filter unit 103b. The third filter unit 103c generates the fourth octave event data V4 of r 3 rows and s 3 columns (r 3 is an integer satisfying r 3 <m 3 , s 3 is an integer satisfying s 3 <n 3 ) from the third octave event data V3 of m 3 rows and n 3 columns (n 3 is an integer satisfying 2 or more, n 3 is an integer satisfying 2 or more). The third filter unit 103c outputs the fourth octave event data V4 of r 3 rows and s 3 columns to the event synthesis unit 104. FIG. 3 shows the fourth octave event data V4 as a plurality of regions P4 arranged in a row. FIG. 3 shows an example where m 3 =2, n 3 =2, r 3 =1, and s 3 =1. Therefore, the region P3 of 2 rows and 2 columns is replaced with a region P4 of 1 row and 1 column. The event data V4 of the fourth octave is also called the event firing presence/absence column of the fourth octave.
 このように、イベント生成部103は、第jオクターブ(jは1≦j≦i-1を満たす整数)のイベントデータVjから、第j+1オクターブのイベントデータVj+1を生成する。これにより、第2~第iオクターブのイベントデータV2~Viを、第1オクターブのイベントデータV1から順次生成することが可能となる。なお、本実施形態の「オクターブ」という用語は、第1~第iオクターブのイベントデータV1~Viの違いが周波数の違いと対応しているため、音楽用語「オクターブ」からの類推で使用されている。 In this way, the event generating unit 103 generates event data Vj+1 of the j+1th octave from event data Vj of the jth octave (j is an integer satisfying 1≦j≦i-1). This makes it possible to sequentially generate event data V2 to Vi of the second to i-th octaves from event data V1 of the first octave. Note that the term "octave" in this embodiment is used by analogy with the musical term "octave" because the differences between the event data V1 to Vi of the first to i-th octaves correspond to differences in frequency.
 [イベント合成部104]
 イベント合成部104は、イベント取得部102から第1オクターブのイベントデータV1を取得し、イベント生成部103から第2~第iオクターブのイベントデータV2~Viを取得する。図3は、i=4の例、すなわち、イベント合成部104が第1~第4オクターブのイベントデータV1~V4を取得する例を示している。
[Event synthesis unit 104]
The event synthesis unit 104 acquires the event data V1 of the first octave from the event acquisition unit 102, and acquires the event data V2 to Vi of the second to i-th octaves from the event generation unit 103. Fig. 3 shows an example where i = 4, that is, an example where the event synthesis unit 104 acquires the event data V1 to V4 of the first to fourth octaves.
 オクターブ情報付加部104aは、第1~第iオクターブのイベントデータV1~Viにオクターブ情報を付加して保持する。オクターブ情報は、第1~第iオクターブのイベントデータV1~Viの識別情報である。本実施形態のオクターブ情報は、イベントデータのオクターブ数に基づく情報であり、例えば、イベントデータのオクターブ数から1を引いた値である。よって、第1~第iオクターブのイベントデータV1~Viのオクターブ情報はそれぞれ「0~i-1」となる。具体的には、第1オクターブのイベントデータV1のオクターブ情報は「0」となり、第2オクターブのイベントデータV2のオクターブ情報は「1」となり、第3オクターブのイベントデータV3のオクターブ情報は「2」となる。 The octave information adding unit 104a adds octave information to the event data V1 to Vi of the first to i-th octaves and stores the added information. The octave information is identification information for the event data V1 to Vi of the first to i-th octaves. In this embodiment, the octave information is based on the number of octaves of the event data, and is, for example, a value obtained by subtracting 1 from the number of octaves of the event data. Therefore, the octave information for the event data V1 to Vi of the first to i-th octaves is "0 to i-1", respectively. Specifically, the octave information for the event data V1 of the first octave is "0", the octave information for the event data V2 of the second octave is "1", and the octave information for the event data V3 of the third octave is "2".
 出力タイミング調整部104bは、オクターブ情報付加部104aが保持しているイベントデータを、イベント出力部105に出力するタイミングを調整する。本実施形態の出力タイミング調整部104bは、第1~第iオクターブのイベントデータV1~Viを、オクターブ数の大きいイベントデータからオクターブ数の大きいイベントデータの順で、イベント出力部105に順番に出力する。この際、第1~第iオクターブのイベントデータV1~Viは、オクターブ情報が付加された状態で出力される。 The output timing adjustment unit 104b adjusts the timing at which the event data held by the octave information addition unit 104a is output to the event output unit 105. In this embodiment, the output timing adjustment unit 104b outputs the event data V1 to Vi of the first to i-th octaves to the event output unit 105 in order from the event data with the largest octave number to the event data with the largest octave number. At this time, the event data V1 to Vi of the first to i-th octaves are output with octave information added.
 [イベント出力部105]
 イベント出力部105は、第1~第iオクターブのイベントデータV1~Viのうちの少なくとも一部のイベントデータを、固体撮像装置100の外部に出力する。イベント出力部105は、イベント合成部104から取得したイベントデータを出力する。本実施形態のイベント出力部105は、図1に示す車両制御システム11にイベントデータを出力するが、代わりに図10または図11に示す電子機器200にイベントデータを出力してもよい。
[Event output unit 105]
The event output unit 105 outputs at least a part of the event data V1 to Vi of the first to i-th octaves to the outside of the solid-state imaging device 100. The event output unit 105 outputs the event data acquired from the event synthesis unit 104. The event output unit 105 of the present embodiment outputs the event data to the vehicle control system 11 shown in Fig. 1, but may instead output the event data to the electronic device 200 shown in Fig. 10 or 11.
 イベントデータ選択部105aは、第1~第iオクターブのイベントデータV1~Viのうち、イベントデータ形成部105bに出力するイベントデータを選択する。例えば、第1~第iオクターブのイベントデータV1~Viから第1~第3オクターブのイベントデータV1~V3が選択された場合には、第1~第3オクターブのイベントデータV1~V3がイベントデータ形成部105bに出力される。 The event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves. For example, when the event data V1 to V3 of the first to third octaves are selected from the event data V1 to Vi of the first to i-th octaves, the event data V1 to V3 of the first to third octaves are output to the event data formation unit 105b.
 イベントデータ形成部105bは、イベントデータ選択部105aにより選択されたイベントデータを、イベント出力データ形式に変換する。イベントデータ形成部105bはその後、イベント出力データ形式に変換されたイベントデータを、オクターブ情報が付加された状態で、固体撮像装置100の外部に出力する。 The event data forming unit 105b converts the event data selected by the event data selecting unit 105a into an event output data format. The event data forming unit 105b then outputs the event data converted into the event output data format to the outside of the solid-state imaging device 100 with the octave information added.
 図4は、第1実施形態の画素アレイ101について説明するための図である。 FIG. 4 is a diagram for explaining the pixel array 101 of the first embodiment.
 図4のA~Cの各々は、イベントデータを出力する画素アレイ101と、このイベントデータを画像表現に変換して得られた画像Eとを示している。画像Eは、文字「A」が新たに出現した状況を撮影した場合の画像に相当する。また、画像Eは、第1オクターブのイベントデータV1から得られた画像に相当する。 Each of A to C in Figure 4 shows a pixel array 101 that outputs event data, and an image E obtained by converting this event data into an image representation. Image E corresponds to an image captured when a situation in which the letter "A" newly appears. Image E also corresponds to an image obtained from event data V1 of the first octave.
 図4のA~Cに示す(x、y、t、p)、(x、y、t、p)、および(x、y、t、p)は、各画素101aのイベントデータを表す。xおよびyは、各画素101aの座標を表す。tは、イベントデータが得られた時刻を表す。pは、イベントデータのポラリティを表す。例えば、オンイベントが発火した際のポラリティは「+」となり、オフイベントが発火した際のポラリティは「-」となる。以上は、他のx、y、t、およびpについても同様である。 ( x0 , y0 , t0 , p0 ), ( x1 , y1 , t1 , p1 ), and ( x2 , y2 , t2 , p2 ) shown in Figures 4A to 4C represent the event data of each pixel 101a. x0 and y0 represent the coordinates of each pixel 101a. t0 represents the time when the event data was obtained. p0 represents the polarity of the event data. For example, the polarity when an on-event is fired is "+", and the polarity when an off-event is fired is "-". The above also applies to the other x, y, t, and p.
 図4のAは、画素数が少なく、解像度が低い画素アレイ101を示している。図4のBは、画素数が中程度で、解像度が中程度の画素アレイ101を示している。図4のCは、画素数が多く、解像度が高い画素アレイ101を示している。 A in FIG. 4 shows a pixel array 101 with a small number of pixels and low resolution. B in FIG. 4 shows a pixel array 101 with a medium number of pixels and medium resolution. C in FIG. 4 shows a pixel array 101 with a large number of pixels and high resolution.
 固体撮像装置100の性能を高めようとして固体撮像装置100を高解像度化すると、データマイニングコストが増大し、固体撮像装置100から出力されたイベントデータを利用する情報処理システムの情報処理が遅延してしまう。その結果、情報処理システム全体の性能はむしろ低下してしまう。このような情報処理システムの例は、図1に示す車両制御システム11や、図10に示す電子機器200や、図11に示すシステム(電子機器200を含むシステム)である。 If the resolution of the solid-state imaging device 100 is increased in an attempt to improve the performance of the solid-state imaging device 100, the data mining costs will increase and the information processing of the information processing system that uses the event data output from the solid-state imaging device 100 will be delayed. As a result, the performance of the entire information processing system will actually decrease. Examples of such information processing systems are the vehicle control system 11 shown in FIG. 1, the electronic device 200 shown in FIG. 10, and the system shown in FIG. 11 (a system including the electronic device 200).
 そこで、本実施形態の固体撮像装置100は、図3を参照して説明したように、様々なオクターブのイベントデータを出力する。これにより、イベントデータを利用した情報処理の遅延を抑制しつつ、固体撮像装置100の高解像度化を実現することが可能となる。例えば、図4のCに示す画素アレイ101を採用しても、イベントデータを利用した情報処理の遅延を抑制することが可能となる。このような効果のさらなる詳細については、後述する。 The solid-state imaging device 100 of this embodiment therefore outputs event data of various octaves, as described with reference to FIG. 3. This makes it possible to achieve high resolution for the solid-state imaging device 100 while suppressing delays in information processing using the event data. For example, even if the pixel array 101 shown in FIG. 4C is employed, it is possible to suppress delays in information processing using the event data. Further details of such effects will be described later.
 図5は、第1実施形態の固体撮像装置100の動作について説明するための図である。 FIG. 5 is a diagram for explaining the operation of the solid-state imaging device 100 of the first embodiment.
 図5の矢印A1は、2行2列分の第1オクターブのイベントデータV1から、1行1列分の第2オクターブのイベントデータV2を生成する処理を示している。この処理では、2行2列分の領域P1が、1行1列分の領域P2に置換されている。図5の矢印A2は、2行2列分の第2オクターブのイベントデータV2から、1行1列分の第3オクターブのイベントデータV3を生成する処理を示している。この処理では、2行2列分の領域P2が、1行1列分の領域P3に置換されている。領域P1~P3内のチェックの有無は、イベントの発火の有無を示している。 Arrow A1 in Figure 5 indicates the process of generating second octave event data V2 of one row and one column from first octave event data V1 of two rows and two columns. In this process, area P1 of two rows and two columns is replaced with area P2 of one row and one column. Arrow A2 in Figure 5 indicates the process of generating third octave event data V3 of one row and one column from second octave event data V2 of two rows and two columns. In this process, area P2 of two rows and two columns is replaced with area P3 of one row and one column. The presence or absence of a check in areas P1 to P3 indicates whether an event has been fired or not.
 図5はさらに、第1オクターブのイベントデータV1を画像表現に変換して得られた画像E1と、第2オクターブのイベントデータV2を画像表現に変換して得られた画像E2と、第3オクターブのイベントデータV3を画像表現に変換して得られた画像E3とを示している。画像E1、E2、E3はそれぞれ、図4のC、B、Aに示す画像Eと同様の画像となっている。このように、本実施形態によれば、第1オクターブのイベントデータV1から第2~第iオクターブのイベントデータV2~Viを生成することで、高解像度の画像から低解像度の画像を生成することが可能となる。後述するように、イベント出力部105は、イベントデータをオクターブごとに画像表現で出力する。これにより、画像表現で出力されたイベントデータを用いることで、イベントデータを画像E1~E3のような態様でオクターブごとに表示することが可能となる。 5 further shows an image E1 obtained by converting the event data V1 of the first octave into an image representation, an image E2 obtained by converting the event data V2 of the second octave into an image representation, and an image E3 obtained by converting the event data V3 of the third octave into an image representation. The images E1, E2, and E3 are similar to the image E shown in C, B, and A of FIG. 4, respectively. Thus, according to this embodiment, it is possible to generate a low-resolution image from a high-resolution image by generating event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave. As will be described later, the event output unit 105 outputs the event data in an image representation for each octave. As a result, it is possible to display the event data for each octave in a form such as images E1 to E3 by using the event data output in an image representation.
 本実施形態のイベントデータは、オクターブ情報が付加されることで、(oc、x、y、t、p)の形で表される。xおよびyは、領域P1~P3に対応する画素101aの座標を表し、例えば、領域P1に対応する1つの画素101aの座標や、領域P2に対応する4個の画素101aの平均座標や、領域P3に対応する16個の画素101aの平均座標を表す。tは、イベントデータが得られた時刻を表す。pは、イベントデータのポラリティを表す。ocは、イベントデータのオクターブ情報を表す。例えば、第1オクターブのイベントデータV1のオクターブ情報ocは「0」となり、第2オクターブのイベントデータV2のオクターブ情報ocは「1」となり、第3オクターブのイベントデータV3のオクターブ情報ocは「2」となる。 In this embodiment, the event data is expressed in the form of (oc, x, y, t, p) by adding octave information. x and y represent the coordinates of the pixels 101a corresponding to the regions P1 to P3, for example, the coordinates of one pixel 101a corresponding to the region P1, the average coordinates of four pixels 101a corresponding to the region P2, or the average coordinates of 16 pixels 101a corresponding to the region P3. t represents the time when the event data was obtained. p represents the polarity of the event data. oc represents the octave information of the event data. For example, the octave information oc of the event data V1 in the first octave is "0", the octave information oc of the event data V2 in the second octave is "1", and the octave information oc of the event data V3 in the third octave is "2".
 図6は、第1実施形態のイベント生成部103の動作を説明するための図である。 FIG. 6 is a diagram for explaining the operation of the event generation unit 103 in the first embodiment.
 図6のAは、m行n列分の第1オクターブのイベントデータV1から、r行s列分の第2オクターブのイベントデータV2を生成する処理を示している。図6のAは、m=2、n=2、r=1、s=1の例を示している。そのため、2行2列分の第1オクターブのイベントデータV1から、1行1列分の第2オクターブのイベントデータV2が生成され、2×2個の領域P1が、1×1個の領域P2に置換されている。同様に、図6のAに示す例では、2行2列分の第jオクターブのイベントデータVjから、1行1列分の第j+1オクターブのイベントデータVj+1が生成される。 A in FIG. 6 shows the process of generating r rows and s columns of second octave event data V2 from m rows and n columns of first octave event data V1. A in FIG. 6 shows an example where m=2, n=2, r=1, s=1. Therefore, 1 row and 1 column of second octave event data V2 is generated from 2 rows and 2 columns of first octave event data V1, and 2×2 regions P1 are replaced with 1×1 regions P2. Similarly, in the example shown in A in FIG. 6, 1 row and 1 column of j+1th octave event data Vj+1 is generated from 2 rows and 2 columns of jth octave event data Vj.
 図6のAにて、kは、m×n個の領域P1に含まれるイベント発火の個数を表す(kは1≦k≦m×nを満たす整数)。例えば、2×2個の領域P1に含まれるイベント発火の個数が3個の場合、kの値は3となる。この場合、2×2個の領域P1のうちの3つは、図5に示す「チェック有」の領域P1に相当しており、2×2個の領域P1のうちの残り1つは、図5に示す「チェック無」の領域P1に相当している。これは、4つの画素101aのうちの3つでイベントが発火し、4つの画素101aのうちの残り1つでイベントが発火しなかったことを示している。 In A of FIG. 6, k represents the number of event firings contained in the m×n regions P1 (k is an integer satisfying 1≦k≦m×n). For example, if the number of event firings contained in the 2×2 regions P1 is 3, the value of k is 3. In this case, three of the 2×2 regions P1 correspond to the "checked" region P1 shown in FIG. 5, and the remaining one of the 2×2 regions P1 corresponds to the "unchecked" region P1 shown in FIG. 5. This indicates that events were fired in three of the four pixels 101a, and no event was fired in the remaining one of the four pixels 101a.
 図6のAに示す例では、4個の領域P1のkが1、2、3、および4のいずれかの場合には、1個の領域P2が「チェック有」となり(図6のC)、4個の領域P1のkが0の場合には、1個の領域P2が「チェック無」となる(図6のB)。すなわち、4個の領域P1のいずれかでイベントが発火した場合には、1個の領域P2でイベントが発火したとして取り扱う。一方、4個の領域P1のいずれでもイベントが発火しなかった場合には、1個の領域P2でイベントが発火しなかったとして取り扱う。これにより、イベント発火の有無に関する情報を、第1オクターブのイベントデータV1から、第2オクターブのイベントデータV2に反映させることが可能となる。これは、第jオクターブのイベントデータVjから、第j+1オクターブのイベントデータVj+1を生成する際にも同様である。図6のBおよびCでは、チェック有の領域P1、P2が黒色で示され、チェック無の領域P1、P2が白色で示されている。 In the example shown in FIG. 6A, when k of the four regions P1 is 1, 2, 3, or 4, one region P2 is "checked" (FIG. 6C), and when k of the four regions P1 is 0, one region P2 is "unchecked" (FIG. 6B). That is, when an event is fired in any of the four regions P1, it is treated as if an event is fired in one region P2. On the other hand, when an event is not fired in any of the four regions P1, it is treated as if an event is not fired in one region P2. This makes it possible to reflect information regarding the presence or absence of an event firing from the event data V1 of the first octave to the event data V2 of the second octave. This is also the case when generating the event data Vj+1 of the j+1th octave from the event data Vj of the jth octave. In FIG. 6B and C, the checked regions P1 and P2 are shown in black, and the unchecked regions P1 and P2 are shown in white.
 図7は、第1実施形態のイベント生成部103の動作を説明するための別の図である。 FIG. 7 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
 図7のA~Cはそれぞれ、図6のA~Cに対応している。ただし、図7のAに示す例では、4個の領域P1のkが2、3、および4のいずれかの場合に、1個の領域P2が「チェック有」となり(図7のC)、4個の領域P1のkが0および1のいずれかの場合に、1個の領域P2が「チェック無」となる(図7のB)。すなわち、4個の領域P1の2個以上でイベントが発火した場合には、1個の領域P2でイベントが発火したとして取り扱う。一方、その他の場合には、1個の領域P2でイベントが発火しなかったとして取り扱う。これは、第jオクターブのイベントデータVjから、第j+1オクターブのイベントデータVj+1を生成する際にも同様である。 A to C in FIG. 7 correspond to A to C in FIG. 6, respectively. However, in the example shown in A in FIG. 7, when k of the four regions P1 is 2, 3, or 4, one region P2 is "checked" (C in FIG. 7), and when k of the four regions P1 is 0 or 1, one region P2 is "unchecked" (B in FIG. 7). In other words, when an event is fired in two or more of the four regions P1, it is treated as if an event was fired in one region P2. On the other hand, in other cases, it is treated as if an event was not fired in one region P2. The same is true when generating event data Vj+1 in the j+1th octave from event data Vj in the jth octave.
 4個の領域P1の1個でイベントが発火した場合、これはノイズイベントの影響である可能性が高い。よって、4個の領域P1の1個でイベントが発火した場合に、1個の領域P2でイベントが発火しなかったとして取り扱うことで、ノイズイベントの影響を抑制することが可能となる。 If an event fires in one of the four regions P1, this is likely due to the influence of a noise event. Therefore, if an event fires in one of the four regions P1, it can be treated as if no event fires in one region P2, making it possible to suppress the influence of the noise event.
 図8は、第1実施形態のイベント生成部103の動作を説明するための別の図である。 FIG. 8 is another diagram for explaining the operation of the event generating unit 103 in the first embodiment.
 図8のA~Cもそれぞれ、図6のA~Cに対応している。ただし、図8のAに示す例では、4個の領域P1のkが2および3のいずれかの場合に、1個の領域P2が「チェック有」となり(図8のC)、4個の領域P1のkが0、1、および4のいずれかの場合に、1個の領域P2が「チェック無」となる(図8のB)。すなわち、4個の領域P1の2個または3個でイベントが発火した場合には、1個の領域P2でイベントが発火したとして取り扱う。一方、その他の場合には、1個の領域P2でイベントが発火しなかったとして取り扱う。これは、第jオクターブのイベントデータVjから、第j+1オクターブのイベントデータVj+1を生成する際にも同様である。 A to C in FIG. 8 correspond to A to C in FIG. 6, respectively. However, in the example shown in A in FIG. 8, when k of the four regions P1 is either 2 or 3, one region P2 is "checked" (C in FIG. 8), and when k of the four regions P1 is either 0, 1, or 4, one region P2 is "unchecked" (B in FIG. 8). In other words, when an event is fired in two or three of the four regions P1, it is treated as if an event was fired in one region P2. On the other hand, in other cases, it is treated as if an event was not fired in one region P2. This is also the case when generating event data Vj+1 in the j+1th octave from event data Vj in the jth octave.
 4個の領域P1の1個でイベントが発火した場合、これはノイズイベントの影響である可能性が高い。また、4個の領域P1の4個でイベントが発火した場合、これはフリッカーイベントの影響である可能性が高い。よって、4個の領域P1の1個または4個でイベントが発火した場合に、1個の領域P2でイベントが発火しなかったとして取り扱うことで、ノイズイベントやフリッカーノイズの影響を抑制することが可能となる。 If an event fires in one of the four regions P1, this is likely to be the result of a noise event. Also, if an event fires in four of the four regions P1, this is likely to be the result of a flicker event. Therefore, if an event fires in one or four of the four regions P1, it is treated as if no event fires in the single region P2, making it possible to suppress the effects of noise events and flicker noise.
 なお、図8のAに示す例では、4個の領域P1のkが1、2、および3のいずれかの場合に、1個の領域P2を「チェック有」とし、4個の領域P1のkが0および4のいずれかの場合に、1個の領域P2を「チェック無」としてもよい。これにより、フリッカーノイズの影響を抑制することが可能となる。 In the example shown in FIG. 8A, when k of the four regions P1 is 1, 2, or 3, one region P2 may be "checked," and when k of the four regions P1 is 0 or 4, one region P2 may be "unchecked." This makes it possible to suppress the effects of flicker noise.
 図9は、第1実施形態のイベント出力部105の動作を説明するための図である。 FIG. 9 is a diagram for explaining the operation of the event output unit 105 in the first embodiment.
 図9のAは、イベントデータ選択部105aの動作を説明するための図である。イベントデータ選択部105aは、第1~第iオクターブのイベントデータV1~Viのうち、イベントデータ形成部105bに出力するイベントデータを選択する。この際、イベントデータ選択部105aは、すべてのオクターブのイベントデータV1~Viを選択してもよいし、一部のオクターブのイベントデータV1~Viのみを選択してもよい。例えば、第1~第iオクターブのイベントデータV1~Viから第1~第3オクターブのイベントデータV1~V3が選択された場合には、第1~第3オクターブのイベントデータV1~V3がイベントデータ形成部105bに出力される。第1~第3オクターブのイベントデータV1~V3はそれぞれ、オクターブ情報「0~2」が付加された状態で出力される。 A in FIG. 9 is a diagram for explaining the operation of the event data selection unit 105a. The event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves. In this case, the event data selection unit 105a may select the event data V1 to Vi of all octaves, or may select only the event data V1 to Vi of some octaves. For example, when the event data V1 to V3 of the first to third octaves are selected from the event data V1 to Vi of the first to i-th octaves, the event data V1 to V3 of the first to third octaves are output to the event data formation unit 105b. The event data V1 to V3 of the first to third octaves are output with the octave information "0 to 2" added, respectively.
 図9のBは、イベントデータ形成部105bの動作を説明するための図である。イベントデータ形成部105bは、イベントデータ選択部105aにより選択されたイベントデータを、イベント出力データ形式に変換する。例えば、イベント出力部105は、イベントデータ選択部105aにより選択されたイベントデータを、オクターブごとにアドレスイベント表現や画像表現に変換する。イベントデータ形成部105bはその後、イベント出力データ形式に変換されたイベントデータを、オクターブ情報が付加された状態で、固体撮像装置100の外部に出力する。固体撮像装置100の外部でイベントデータを用いた情報処理を行う際には、画像表現で出力されたイベントデータを用いることで、イベントデータを画像E1~E3のような態様でオクターブごとに表示することが可能となる。 B of FIG. 9 is a diagram for explaining the operation of the event data formation unit 105b. The event data formation unit 105b converts the event data selected by the event data selection unit 105a into an event output data format. For example, the event output unit 105 converts the event data selected by the event data selection unit 105a into an address event representation or an image representation for each octave. The event data formation unit 105b then outputs the event data converted into the event output data format to the outside of the solid-state imaging device 100 with octave information added. When performing information processing using the event data outside the solid-state imaging device 100, the event data output in the image representation can be used to display the event data for each octave in a format such as images E1 to E3.
 (3)第1実施形態の電子機器200
 図10は、第1実施形態の電子機器200の一例を示す図である。
(3) Electronic Device 200 of First Embodiment
FIG. 10 is a diagram illustrating an example of the electronic device 200 according to the first embodiment.
 図10のAは、電子機器200の一例として、カメラ機能を有するスマートフォンを示している。この電子機器200は、図10のBに示すように、撮像部201と、表示部202と、情報処理部203と、記憶部204と、入力部205とを備えている。情報処理部203は、抽出部203aと、選択部203bとを含んでいる。 A in FIG. 10 shows a smartphone with a camera function as an example of electronic device 200. As shown in B in FIG. 10, this electronic device 200 includes an imaging unit 201, a display unit 202, an information processing unit 203, a storage unit 204, and an input unit 205. The information processing unit 203 includes an extraction unit 203a and a selection unit 203b.
 撮像部201は、カメラ機能を実現するための機能ブロックである。撮像部201は、図3に示す固体撮像装置100を含んでいる。この電子機器200は、固体撮像装置100(イベント出力部105)から出力されたイベントデータを利用した情報処理を行う情報処理システムとして機能する。 The imaging unit 201 is a functional block for implementing a camera function. The imaging unit 201 includes the solid-state imaging device 100 shown in FIG. 3. This electronic device 200 functions as an information processing system that performs information processing using event data output from the solid-state imaging device 100 (event output unit 105).
 表示部202は、文字や画像を表示するための表示画面を有する。例えば、表示部202は、固体撮像装置100から出力されたイベントデータを表示画面に表示する。本実施形態では、イベント出力部105が、イベントデータを画像表現で出力し、表示部202が、このイベントデータを画像の形で表示画面に表示する(図10のB)。この画像は、静止画でも動画でもよい。図10のAでは、表示画面が、イメージセンサとしての固体撮像装置100により撮像された画像と、イベントセンサ(EVS)としての固体撮像装置100から出力されたイベントデータを画像の形で表示した画像とを表示している。表示画面にイベントデータを表示するモードでは、表示画面はビューワーとも呼ばれる。 The display unit 202 has a display screen for displaying characters and images. For example, the display unit 202 displays event data output from the solid-state imaging device 100 on the display screen. In this embodiment, the event output unit 105 outputs the event data in an image representation, and the display unit 202 displays this event data in the form of an image on the display screen (B in FIG. 10). This image may be a still image or a moving image. In A in FIG. 10, the display screen displays an image captured by the solid-state imaging device 100 as an image sensor, and an image in which the event data output from the solid-state imaging device 100 as an event sensor (EVS) is displayed in the form of an image. In a mode in which event data is displayed on the display screen, the display screen is also called a viewer.
 情報処理部203は、電子機器200の制御などの種々の情報処理を行う。例えば、情報処理部203は、固体撮像装置100からイベントデータを受信し、このイベントデータを表示部202の表示画面に表示する。 The information processing unit 203 performs various information processes such as controlling the electronic device 200. For example, the information processing unit 203 receives event data from the solid-state imaging device 100 and displays the event data on the display screen of the display unit 202.
 記憶部204は、半導体メモリなどの記録媒体を含んでいる。情報処理部203は、情報処理に必要な情報を記憶部204から読み出すことや、情報処理により生じた情報を記憶部204内に記録することができる。例えば、情報処理部203は、固体撮像装置100からイベントデータを受信し、このイベントデータを記憶部204内に記録する。 The storage unit 204 includes a recording medium such as a semiconductor memory. The information processing unit 203 can read information necessary for information processing from the storage unit 204, and record information generated by information processing in the storage unit 204. For example, the information processing unit 203 receives event data from the solid-state imaging device 100, and records this event data in the storage unit 204.
 入力部205は、ユーザーからの入力操作を受け付ける。情報処理部203は、この入力操作に応じた情報処理を行う。入力部205は例えば、タッチパネルやハードボタンを含んでいる。 The input unit 205 accepts input operations from the user. The information processing unit 203 performs information processing according to the input operations. The input unit 205 includes, for example, a touch panel and hard buttons.
 イベント出力部105では、イベントデータ選択部105aが、第1~第iオクターブのイベントデータV1~Viのうち、イベントデータ形成部105bに出力するイベントデータを選択する。例えば、第1~第iオクターブのイベントデータV1~Viから第1~第3オクターブのイベントデータV1~V3が選択された場合には、第1~第3オクターブのイベントデータV1~V3がイベントデータ形成部105bに出力される。イベントデータ形成部105bは、イベントデータ選択部105aにより選択されたイベントデータをオクターブごとに画像表現で出力する。例えば、第1~第3オクターブのイベントデータV1~V3が、オクターブごとに画像表現で出力される。 In the event output unit 105, the event data selection unit 105a selects the event data to be output to the event data formation unit 105b from the event data V1 to Vi of the first to i-th octaves. For example, when the event data V1 to V3 of the first to third octaves is selected from the event data V1 to Vi of the first to i-th octaves, the event data V1 to V3 of the first to third octaves is output to the event data formation unit 105b. The event data formation unit 105b outputs the event data selected by the event data selection unit 105a in an image representation for each octave. For example, the event data V1 to V3 of the first to third octaves is output in an image representation for each octave.
 抽出部203aは、固体撮像装置100(イベントデータ形成部105b)から出力されたイベントデータから、所定のオクターブ数のイベントデータを抽出する。例えば、抽出部203aは、第1~第3オクターブのイベントデータV1~V3から、第2オクターブのイベントデータV2を抽出する。所定のオクターブ数のイベントデータは、イベントデータのオクターブ情報に基づいて抽出可能である。情報処理部203は、抽出部203aにより抽出されたイベントデータを表示画面に表示する。例えば、第2オクターブのイベントデータV2が抽出された場合には、第2オクターブのイベントデータV2が画像の形で表示画面に表示される。 The extraction unit 203a extracts event data of a predetermined number of octaves from the event data output from the solid-state imaging device 100 (event data formation unit 105b). For example, the extraction unit 203a extracts event data V2 of the second octave from event data V1 to V3 of the first to third octaves. Event data of a predetermined number of octaves can be extracted based on the octave information of the event data. The information processing unit 203 displays the event data extracted by the extraction unit 203a on the display screen. For example, when event data V2 of the second octave is extracted, the event data V2 of the second octave is displayed on the display screen in the form of an image.
 抽出部203aは、ビューワーの解像度に合うオクターブ数のイベントデータを自動的に抽出する。例えば、ビューワーに最初にイベントデータを表示する際には、第2オクターブのイベントデータV2が抽出され表示される。その後、ユーザーがビューワーの解像度を上げる操作を行うと、第1オクターブのイベントデータV1が抽出され表示される。一方、ユーザーがビューワーの解像度を下げる操作を行うと、第3オクターブのイベントデータV3が抽出され表示される。 The extraction unit 203a automatically extracts event data for the number of octaves that match the resolution of the viewer. For example, when event data is first displayed in the viewer, event data V2 of the second octave is extracted and displayed. Thereafter, when the user performs an operation to increase the resolution of the viewer, event data V1 of the first octave is extracted and displayed. On the other hand, when the user performs an operation to decrease the resolution of the viewer, event data V3 of the third octave is extracted and displayed.
 選択部203bは、固体撮像装置100(イベントデータ形成部105b)から出力されたイベントデータから、ユーザーにより指定されたオクターブ数のイベントデータを選択する。例えば、第2オクターブのイベントデータV2がビューワーに表示されている場合に、ユーザーがタッチパネル上の「記録開始ボタン」にタッチすると、選択部203bは、第1~第3オクターブのイベントデータV1~V3から、第2オクターブのイベントデータV2を選択する。そして、選択部203bは、第2オクターブのイベントデータV2を記憶部204内に記録し始める。その後、ユーザーがタッチパネル上の「記録終了ボタン」にタッチすると、第2オクターブのイベントデータV2の記録を終了する。このようにして、ユーザーは、記録するイベントデータのオクターブ数を指定することや、記録を開始および終了するタイミングを指定することができる。その結果、記録開始から記録終了までの動画が、記憶部204(記録媒体)内に記録される。 The selection unit 203b selects the event data for the number of octaves specified by the user from the event data output from the solid-state imaging device 100 (event data formation unit 105b). For example, when the event data V2 for the second octave is displayed on the viewer, if the user touches the "recording start button" on the touch panel, the selection unit 203b selects the event data V2 for the second octave from the event data V1 to V3 for the first to third octaves. Then, the selection unit 203b starts recording the event data V2 for the second octave in the storage unit 204. After that, if the user touches the "recording end button" on the touch panel, the recording of the event data V2 for the second octave ends. In this way, the user can specify the number of octaves of the event data to be recorded and the timing to start and end recording. As a result, a video from the start of recording to the end of recording is recorded in the storage unit 204 (recording medium).
 なお、記憶部204に記録されるイベントデータのオクターブ数は、ビューワーに表示されているイベントデータのオクターブ数と異なっていてもよい。例えば、第2オクターブのイベントデータV2がビューワーに表示されている場合に、ユーザーがタッチパネル上で「記録対象」として「第1オクターブのイベントデータV1」を指定すると、第1オクターブのイベントデータV1が記憶部204内に記録されてもよい。また、このような指定がなくても、第1オクターブのイベントデータV1が記憶部204内に記録されてもよい。 The number of octaves of the event data recorded in the memory unit 204 may be different from the number of octaves of the event data displayed in the viewer. For example, when the event data V2 of the second octave is displayed in the viewer, if the user specifies "event data V1 of the first octave" as "to be recorded" on the touch panel, the event data V1 of the first octave may be recorded in the memory unit 204. Furthermore, even if no such specification is made, the event data V1 of the first octave may be recorded in the memory unit 204.
 また、情報処理部203は、抽出されたイベントデータを、表示以外の情報処理に利用してもよい。例えば、第2オクターブのイベントデータV2が抽出された後に、情報処理部203は、第2オクターブのイベントデータV2を画像認識に利用してもよい。画像認識の例は、ユーザーのジェスチャーの認識である。例えば、情報処理部203は、第2オクターブのイベントデータV2に含まれるユーザーの画像を用いて、ユーザーのジェスチャーを認識するための画像認識を行ってもよい。この際、情報処理部203は、複数のオクターブ数のイベントデータを画像認識に利用してもよい。 In addition, the information processing unit 203 may use the extracted event data for information processing other than display. For example, after the event data V2 of the second octave is extracted, the information processing unit 203 may use the event data V2 of the second octave for image recognition. An example of image recognition is recognition of a user's gesture. For example, the information processing unit 203 may perform image recognition to recognize a user's gesture using an image of the user included in the event data V2 of the second octave. At this time, the information processing unit 203 may use event data of multiple octaves for image recognition.
 本実施形態では、固体撮像装置100が様々なオクターブ数のイベントデータを出力する。よって、情報処理部203は、出力されたイベントデータから所定のオクターブ数のイベントデータを抽出することで、様々なオクターブ数のイベントデータを表示することができる。また、情報処理部203は、出力されたイベントデータから所定のオクターブ数のイベントデータを選択することで、様々なオクターブ数のイベントデータを記録することができる。仮に情報処理部203が、ある解像度のイベントデータから別の解像度のイベントデータを生成することにすると、情報処理部203による情報処理が遅延してしまう。本実施形態によれば、様々な解像度(オクターブ数)のイベントデータを生成する処理を固体撮像装置100に担わせることで、情報処理部203による情報処理の遅延を抑制することが可能となる。本実施形態では、様々な解像度のイベントデータを生成する処理を、情報処理部203によりソフトウェア的に行う代わりに、固体撮像装置100によりハードウェア的に行っている。 In this embodiment, the solid-state imaging device 100 outputs event data of various octaves. Therefore, the information processing unit 203 can display event data of various octaves by extracting event data of a predetermined number of octaves from the output event data. In addition, the information processing unit 203 can record event data of various octaves by selecting event data of a predetermined number of octaves from the output event data. If the information processing unit 203 were to generate event data of a different resolution from event data of a certain resolution, the information processing by the information processing unit 203 would be delayed. According to this embodiment, the solid-state imaging device 100 is responsible for the process of generating event data of various resolutions (number of octaves), so that the delay in information processing by the information processing unit 203 can be suppressed. In this embodiment, the process of generating event data of various resolutions is performed by the solid-state imaging device 100 as hardware, instead of by the information processing unit 203 as software.
 図11は、第1実施形態の電子機器200の別の一例を示す図である。 FIG. 11 is a diagram showing another example of the electronic device 200 of the first embodiment.
 図11のAは、電子機器200の一例として、ゲームコンソールを示している。この電子機器200は、図11のAに示すように、撮像装置201’および表示装置202’と有線または無線で接続されて使用される。この電子機器200は、図11のBに示すように、情報処理部203と、記憶部204と、入力部205とを備えている。情報処理部203は、抽出部203aと、選択部203bとを含んでいる。 A in FIG. 11 shows a game console as an example of electronic device 200. As shown in A in FIG. 11, this electronic device 200 is used by being connected to an imaging device 201' and a display device 202' via wired or wireless connection. As shown in B in FIG. 11, this electronic device 200 includes an information processing unit 203, a storage unit 204, and an input unit 205. The information processing unit 203 includes an extraction unit 203a and a selection unit 203b.
 撮像装置201’は例えば、ゲームコンソールの付属品のカメラである。撮像装置201’は、前述の撮像部201と同様に、図3に示す固体撮像装置100を含んでいる。この電子機器200は、撮像装置201’および表示装置202’と共に、固体撮像装置100(イベント出力部105)から出力されたイベントデータを利用した情報処理を行う情報処理システムを構成している。 The imaging device 201' is, for example, a camera that is an accessory to a game console. The imaging device 201' includes the solid-state imaging device 100 shown in FIG. 3, similar to the imaging unit 201 described above. This electronic device 200, together with the imaging device 201' and the display device 202', constitutes an information processing system that performs information processing using event data output from the solid-state imaging device 100 (event output unit 105).
 表示装置202’は例えば、液晶大画面テレビである。表示装置202’は、前述の表示部202と同様に、文字や画像を表示するための表示画面を有する。図11のAでは、撮像装置201’が、ゲームコンソールで遊んでいるユーザーを撮像している。図11のBでは、表示装置202’が、このユーザーを撮像して得られたイベントデータを画像の形で表示画面に表示している。表示画面にイベントデータを表示するモードでは、表示画面はビューワーとも呼ばれる。 The display device 202' is, for example, a large LCD television. The display device 202' has a display screen for displaying characters and images, similar to the display unit 202 described above. In A of FIG. 11, the imaging device 201' captures an image of a user playing on a game console. In B of FIG. 11, the display device 202' captures an image of the user and displays the event data obtained in the form of an image on the display screen. In the mode in which event data is displayed on the display screen, the display screen is also called a viewer.
 図11のBに示す情報処理部203、記憶部204、および入力部205の機能は、図10のAに示す情報処理部203、記憶部204、および入力部205の機能とおおむね同様である。 The functions of the information processing unit 203, memory unit 204, and input unit 205 shown in FIG. 11B are generally similar to the functions of the information processing unit 203, memory unit 204, and input unit 205 shown in FIG. 10A.
 図12は、図11に示す電子機器200の詳細を説明するための図である。 FIG. 12 is a diagram for explaining the details of the electronic device 200 shown in FIG. 11.
 撮像装置201’は、図11のAに示すように、ユーザーの全身を撮像している。そのため、撮像装置201’(固体撮像装置100)から出力される第1~第3オクターブのイベントデータV1~V3は、ユーザーの全身に関するイベントデータを含んでいる。 As shown in A of FIG. 11, the imaging device 201' captures the entire body of the user. Therefore, the event data V1 to V3 of the first to third octaves output from the imaging device 201' (solid-state imaging device 100) includes event data related to the entire body of the user.
 図12のAでは、表示装置202’の表示画面が、第3オクターブのイベントデータV3を用いて、ユーザーの全身を低解像度で表示している。図12のAにて点線で囲む領域内の画像E3は、ユーザーの全身を含んでいる。第3オクターブのイベントデータV3の全部を表示画面上に表示すると、画像E3が得られる。 In FIG. 12A, the display screen of the display device 202' displays the user's entire body at low resolution using the third octave event data V3. Image E3 in the area enclosed by the dotted line in FIG. 12A includes the user's entire body. When all of the third octave event data V3 is displayed on the display screen, image E3 is obtained.
 図12のBでは、表示装置202’の表示画面が、第2オクターブのイベントデータV2を用いて、ユーザーの全身を中解像度で表示している。図12のBにて点線で囲む領域内の画像E2は、ユーザの手の部分を含んでいる。この画像E2は、第2オクターブのイベントデータV2の一部に相当する。 In FIG. 12B, the display screen of the display device 202' displays the user's entire body at medium resolution using the second octave event data V2. Image E2 in the area enclosed by the dotted line in FIG. 12B includes the user's hands. This image E2 corresponds to a portion of the second octave event data V2.
 図12のCでは、表示装置202’の表示画面が、第1オクターブのイベントデータV1を用いて、ユーザーの手の部分を高解像度で表示している。図12のCにて点線で囲む領域内の画像E1は、ユーザの手の部分を含んでいる。この画像E1は、第1オクターブのイベントデータV1の一部に相当する。 In FIG. 12C, the display screen of the display device 202' displays the user's hand in high resolution using the event data V1 of the first octave. Image E1 in the area enclosed by the dotted line in FIG. 12C includes the user's hand. This image E1 corresponds to a portion of the event data V1 of the first octave.
 情報処理部203は、このような画像E1~E3を利用することで、ユーザーの全身を拡大または縮小して表示することができる。例えば、表示画面の表示内容を、図12のAから図12のBへと遷移させることで、画像の解像度を上げつつ、ユーザーの全身を拡大して表示することができる。また、表示画面の表示内容を、図12のBから図12のCへと遷移させることで、画像の解像度を上げつつ、ユーザーの手の部分を拡大して表示することができる。これにより、ユーザーの手のジェスチャーを、表示画面上で確認することが可能となる。図12のAから図12のBへの遷移は、抽出部203aにより抽出されるイベントデータを、第3オクターブのイベントデータV3から第2オクターブのイベントデータV2に切り替えることで実現可能である。図12のBから図12のCへの遷移は、抽出部203aにより抽出されるイベントデータを、第2オクターブのイベントデータV2から第1オクターブのイベントデータV1に切り替えることで実現可能である。 The information processing unit 203 can use these images E1 to E3 to enlarge or reduce the entire body of the user for display. For example, by transitioning the display content of the display screen from A in FIG. 12 to B in FIG. 12, the entire body of the user can be enlarged and displayed while increasing the image resolution. Also, by transitioning the display content of the display screen from B in FIG. 12 to C in FIG. 12, the hand of the user can be enlarged and displayed while increasing the image resolution. This makes it possible to confirm the gesture of the user's hand on the display screen. The transition from A in FIG. 12 to B in FIG. 12 can be realized by switching the event data extracted by the extraction unit 203a from the event data V3 of the third octave to the event data V2 of the second octave. The transition from B in FIG. 12 to C in FIG. 12 can be realized by switching the event data extracted by the extraction unit 203a from the event data V2 of the second octave to the event data V1 of the first octave.
 情報処理部203は、ジェスチャーの確認を自動化するために、抽出部203aにより抽出されたイベントデータを利用して画像認識を行ってもよい。例えば、情報処理部203は、第3オクターブのイベントデータV3から、画像E3の領域、すなわち、ユーザーの全身を画像認識により抽出する。次に、情報処理部203は、第2オクターブのイベントデータV2から、画像E2の領域、すなわち、ユーザーの手の部分を画像認識により抽出する。次に、情報処理部203は、第1オクターブのイベントデータV1から、ユーザーの手のジェスチャーを画像認識により特定する。これにより、ユーザーの手のジェスチャーを、情報処理部203が自動的に認識することが可能となる。本実施形態によれば、様々な解像度(オクターブ数)のイベントデータを生成する処理を固体撮像装置100に担わせることで、このような画像認識の遅延を抑制することが可能となる。 The information processing unit 203 may perform image recognition using the event data extracted by the extraction unit 203a to automate the confirmation of gestures. For example, the information processing unit 203 extracts the area of image E3, i.e., the user's entire body, from the event data V3 of the third octave by image recognition. Next, the information processing unit 203 extracts the area of image E2, i.e., the user's hand part, from the event data V2 of the second octave by image recognition. Next, the information processing unit 203 identifies the user's hand gesture from the event data V1 of the first octave by image recognition. This makes it possible for the information processing unit 203 to automatically recognize the user's hand gesture. According to this embodiment, by having the solid-state imaging device 100 take charge of the process of generating event data of various resolutions (number of octaves), it is possible to suppress such delays in image recognition.
 以上のように、本実施形態の固体撮像装置100は、画素アレイ101から出力されたイベントデータを第1オクターブのイベントデータV1とし、第1オクターブのイベントデータV1から第2~第iオクターブのイベントデータV2~Viを生成する。よって、本実施形態によれば、イベントデータを利用した情報処理用に様々なオクターブのイベントデータを出力することで、情報処理の遅延を抑制しつつ、固体撮像装置100の高解像度化を実現することが可能となる。 As described above, the solid-state imaging device 100 of this embodiment treats the event data output from the pixel array 101 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave. Therefore, according to this embodiment, by outputting event data of various octaves for information processing using the event data, it is possible to achieve high resolution of the solid-state imaging device 100 while suppressing delays in information processing.
 (第2実施形態)
 図13は、第2実施形態の固体撮像装置100の構成を示すブロック図である。
Second Embodiment
FIG. 13 is a block diagram showing the configuration of a solid-state imaging device 100 according to the second embodiment.
 本実施形態の固体撮像装置100は、第1実施形態の固体撮像装置と同様に、画素アレイ101と、イベント取得部102と、イベント生成部103と、イベント合成部104と、イベント出力部105とを備えている。本実施形態の固体撮像装置100はさらに、フレームメモリ111を備えている。また、本実施形態のイベント取得部102は、アービタ部102aと、タイムスタンプ部102bとを含んでいる。 The solid-state imaging device 100 of this embodiment, like the solid-state imaging device of the first embodiment, includes a pixel array 101, an event acquisition unit 102, an event generation unit 103, an event synthesis unit 104, and an event output unit 105. The solid-state imaging device 100 of this embodiment further includes a frame memory 111. The event acquisition unit 102 of this embodiment also includes an arbiter unit 102a and a time stamp unit 102b.
 本実施形態の固体撮像装置100は、アービタ型EVSである。よって、本実施形態のイベント取得部102は、画素アレイ102内の複数の画素102aからイベントデータをランダムな順番で取得する。イベント取得部102が取得したイベントデータは、一定期間フレームメモリ111内に格納された後、フレームメモリ111からイベント生成部103およびイベント合成部104に出力される。フレームメモリ111は、画素アレイ101と同様に、2次元アレイ状(マトリクス状)に配置された複数のメモリセルを含んでいる。フレームメモリ111の各行のイベントデータは、第1オクターブのイベントデータV1として取り扱われる。 The solid-state imaging device 100 of this embodiment is an arbiter-type EVS. Therefore, the event acquisition unit 102 of this embodiment acquires event data from multiple pixels 102a in the pixel array 102 in a random order. The event data acquired by the event acquisition unit 102 is stored in the frame memory 111 for a certain period of time, and is then output from the frame memory 111 to the event generation unit 103 and the event synthesis unit 104. The frame memory 111 includes multiple memory cells arranged in a two-dimensional array (matrix), similar to the pixel array 101. The event data of each row of the frame memory 111 is treated as event data V1 of the first octave.
 本実施形態のイベント取得部102は、アービタ型EVS用の機能ブロックとして、アービタ部102aとタイムスタンプ部102bとを含んでいる。アービタ部102aは、複数の画素101aから出力された複数のイベント(リクエスト信号)を調停する。タイムスタンプ部102bは、各画素101aから発火したイベントにタイムスタンプを付与する。本実施形態では、各画素101aのイベントデータ(x、y、t、p)におけるtの値が、タイムスタンプの値となる。 The event acquisition unit 102 of this embodiment includes an arbiter unit 102a and a time stamp unit 102b as functional blocks for an arbiter-type EVS. The arbiter unit 102a arbitrates multiple events (request signals) output from multiple pixels 101a. The time stamp unit 102b assigns a time stamp to an event fired from each pixel 101a. In this embodiment, the value of t in the event data (x, y, t, p) of each pixel 101a becomes the value of the time stamp.
 図14は、第2実施形態のイベント生成部103の動作を説明するための図である。 FIG. 14 is a diagram for explaining the operation of the event generation unit 103 in the second embodiment.
 図14は、図6のA、図7のA、および図8のAと同様に、m行n列分の第1オクターブのイベントデータV1から、r行s列分の第2オクターブのイベントデータV2を生成する処理を示している。図14は、m=2、n=2、r=1、s=1の例を示している。そのため、2行2列分の第1オクターブのイベントデータV1から、1行1列分の第2オクターブのイベントデータV2が生成され、2×2個の領域P1が、1×1個の領域P2に置換されている。同様に、図14に示す例では、2行2列分の第jオクターブのイベントデータVjから、1行1列分の第j+1オクターブのイベントデータVj+1が生成される。 FIG. 14, like A of FIG. 6, A of FIG. 7, and A of FIG. 8, shows the process of generating r rows and s columns of second octave event data V2 from m rows and n columns of first octave event data V1. FIG. 14 shows an example where m=2, n=2, r=1, and s=1. Therefore, 1 row and 1 column of second octave event data V2 is generated from 2 rows and 2 columns of first octave event data V1, and 2×2 regions P1 are replaced with 1×1 regions P2. Similarly, in the example shown in FIG. 14, 1 row and 1 column of j+1th octave event data Vj+1 is generated from 2 rows and 2 columns of jth octave event data Vj.
 図14に示す例における処理は、図6のA、図7のA、または図8のAに示す例における処理と同様に行うことが可能である。ただし、本実施形態では、各画素101aのイベントデータ(x、y、t、p)におけるtの値が、タイムスタンプの値となるため、図14に示す4個の領域P1が、互いに異なるtの値を有している。図14では、4個の領域P1のtの値が、それぞれta、tb、tc、およびtdとなっている。 The processing in the example shown in FIG. 14 can be performed in the same manner as the processing in the example shown in FIG. 6A, FIG. 7A, or FIG. 8A. However, in this embodiment, the value of t in the event data (x, y, t, p) of each pixel 101a becomes the value of the timestamp, so the four regions P1 shown in FIG. 14 have different values of t. In FIG. 14, the values of t in the four regions P1 are ta, tb, tc, and td, respectively.
 本実施形態のイベントデータは、第1実施形態のイベントデータと同様に、オクターブ情報が付加されることで、(oc、x、y、t、p)の形で表される。xおよびyは、領域P1~P3等に対応する画素101aの座標を表し、例えば、領域P1に対応する1つの画素101aの座標や、領域P2に対応する4個の画素101aの平均座標や、領域P3に対応する16個の画素101aの平均座標を表す。tは、イベントデータが得られた時刻を表す。pは、イベントデータのポラリティを表す。 The event data of this embodiment, like the event data of the first embodiment, is expressed in the form (oc, x, y, t, p) with the addition of octave information. x and y represent the coordinates of the pixels 101a corresponding to regions P1 to P3, etc., and represent, for example, the coordinates of one pixel 101a corresponding to region P1, the average coordinates of four pixels 101a corresponding to region P2, or the average coordinates of 16 pixels 101a corresponding to region P3. t represents the time when the event data was obtained. p represents the polarity of the event data.
 図14では、1つの領域P2のtを、4個の領域P1のtの統計値とする。すなわち、1つの領域P2のtを、4個の領域P1のtに対し統計的な処理を施して得られる値とする。4個の領域P1のtの統計値は例えば、4個の領域P1のtの平均値、最大値、最小値などである。図14は、1つの領域P2のt’を、4個の領域P1のta~tdの平均値とする例を示している(t’=(ta+tb+tc+td)/4)。同様に、4つの領域Pjを1つの領域Pj+1に置換する際にも、1つの領域Pj+1のtを、4個の領域Pjのtの統計値とする。 In FIG. 14, the t of one region P2 is taken as the statistical value of the t of four regions P1. In other words, the t of one region P2 is taken as the value obtained by performing statistical processing on the t of four regions P1. The statistical value of the t of four regions P1 is, for example, the average value, maximum value, minimum value, etc. of the t of the four regions P1. FIG. 14 shows an example in which the t' of one region P2 is taken as the average value of ta to td of the four regions P1 (t' = (ta + tb + tc + td) / 4). Similarly, when replacing four regions Pj with one region Pj+1, the t of one region Pj+1 is taken as the statistical value of the t of the four regions Pj.
 本実施形態の固体撮像装置100は、フレームメモリ111から出力されたイベントデータを第1オクターブのイベントデータV1とし、第1オクターブのイベントデータV1から第2~第iオクターブのイベントデータV2~Viを生成する。よって、本実施形態によれば、第1実施形態と同様に、イベントデータを利用した情報処理用に様々なオクターブのイベントデータを出力することで、情報処理の遅延を抑制しつつ、固体撮像装置100の高解像度化を実現することが可能となる。 The solid-state imaging device 100 of this embodiment treats the event data output from the frame memory 111 as event data V1 of the first octave, and generates event data V2 to Vi of the second to i-th octaves from the event data V1 of the first octave. Therefore, according to this embodiment, as in the first embodiment, by outputting event data of various octaves for information processing using the event data, it is possible to achieve high resolution of the solid-state imaging device 100 while suppressing delays in information processing.
 (第3実施形態)
 図15は、第3実施形態の固体撮像装置100の構成を模式的に示す斜視図である。
Third Embodiment
FIG. 15 is a perspective view that illustrates a schematic configuration of a solid-state imaging device 100 according to the third embodiment.
 本実施形態の固体撮像装置100は、検出チップ120と、検出チップ120上に積層された受光チップ130とを備えている。検出チップ120と受光チップ130は、ビアプラグ、金属パッド、金属バンプなどの接続部を介して電気的に接続されている。本実施形態の固体撮像装置100は、第1または第2実施形態の固体撮像装置100として機能する。 The solid-state imaging device 100 of this embodiment includes a detection chip 120 and a light-receiving chip 130 stacked on the detection chip 120. The detection chip 120 and the light-receiving chip 130 are electrically connected through connections such as via plugs, metal pads, and metal bumps. The solid-state imaging device 100 of this embodiment functions as the solid-state imaging device 100 of the first or second embodiment.
 図15は、互いに垂直なX軸、Y軸、およびZ軸を示している。X方向およびY方向は横方向に相当し、Z方向は縦方向に相当する。また、+Z方向は上方向に相当し、-Z方向は下方向に相当する。なお、-Z方向は、厳密に重力方向に一致していてもよいし、厳密には重力方向に一致していなくてもよい。 FIG. 15 shows the X-axis, Y-axis, and Z-axis, which are perpendicular to each other. The X-axis and Y-axis correspond to the horizontal direction, and the Z-axis corresponds to the vertical direction. The +Z direction corresponds to the upward direction, and the -Z direction corresponds to the downward direction. Note that the -Z direction may or may not strictly coincide with the direction of gravity.
 図16は、第3実施形態の受光チップ120の構成を模式的に示す平面図である。 FIG. 16 is a plan view that shows a schematic configuration of the light receiving chip 120 of the third embodiment.
 図16のAは、受光チップ120の平面構造の例を示している。受光チップ120は、受光部121と、複数のビア配置部122~124とを含んでいる。図16のBは、受光部121の平面構造の例を示している。受光部121は複数のフォトダイオード121aを含んでいる。 A in FIG. 16 shows an example of the planar structure of the light-receiving chip 120. The light-receiving chip 120 includes a light-receiving section 121 and multiple via arrangement sections 122 to 124. B in FIG. 16 shows an example of the planar structure of the light-receiving section 121. The light-receiving section 121 includes multiple photodiodes 121a.
 受光部121では、複数のフォトダイオード121aがアレイ状(二次元格子状)に配列されている。各フォトダイオード121aには、行アドレスおよび列アドレスからなる画素アドレスが割り当てられており、各フォトダイオード121aは、画素として扱われる。各フォトダイオード121aは、入射光を光電変換して光電流を生成する。ビア配置部122~124には、検出チップ130と電気的に接続されたビアプラグが配置されている。 In the light receiving section 121, multiple photodiodes 121a are arranged in an array (two-dimensional lattice). A pixel address consisting of a row address and a column address is assigned to each photodiode 121a, and each photodiode 121a is treated as a pixel. Each photodiode 121a photoelectrically converts incident light to generate a photocurrent. Via plugs electrically connected to the detection chip 130 are arranged in the via arrangement sections 122 to 124.
 図17は、第3実施形態の検出チップ130の構成を模式的に示す平面図である。 FIG. 17 is a plan view that shows a schematic configuration of the detection chip 130 of the third embodiment.
 図17のAは、検出チップ130の平面構造の例を示している。検出チップ130は、アドレスイベント検出部131と、複数のビア配置部132~134と、行駆動回路135と、列駆動回路136と、信号処理回路137とを含んでいる。図17のBは、アドレスイベント検出部131の平面構造の例を示している。アドレスイベント検出部131は複数のアドレスイベント検出回路131aを含んでいる。 A in FIG. 17 shows an example of the planar structure of the detection chip 130. The detection chip 130 includes an address event detection unit 131, multiple via placement units 132 to 134, a row driving circuit 135, a column driving circuit 136, and a signal processing circuit 137. B in FIG. 17 shows an example of the planar structure of the address event detection unit 131. The address event detection unit 131 includes multiple address event detection circuits 131a.
 アドレスイベント検出部131では、複数のアドレスイベント検出回路131aがアレイ状(二次元格子状)に配列されている。各アドレスイベント検出回路131aには、画素アドレスが割り当てられており、各アドレスイベント検出回路131aは、同ーアドレスのフォトダイオード121aと電気的に接続されている。各アドレスイベント検出回路121aは、対応するフォトダイオード121aからの光電流に応じた電圧信号を量子化して検出信号として出力する。この検出信号は、入射光の光量が所定の閾値を超えた旨をアドレスイベントとして検出したか否かを示す1ビットの信号であり、信号処理回路137に出力される。ビア配置部132~134には、受光チップ120と電気的に接続されたビアプラグが配置されている。 In the address event detection section 131, multiple address event detection circuits 131a are arranged in an array (two-dimensional lattice). A pixel address is assigned to each address event detection circuit 131a, and each address event detection circuit 131a is electrically connected to the photodiode 121a at the same address. Each address event detection circuit 121a quantizes a voltage signal corresponding to the photocurrent from the corresponding photodiode 121a and outputs it as a detection signal. This detection signal is a one-bit signal indicating whether or not an address event has been detected in which the amount of incident light has exceeded a predetermined threshold, and is output to the signal processing circuit 137. Via plugs electrically connected to the light receiving chip 120 are arranged in the via arrangement sections 132 to 134.
 行駆動回路135は、行アドレスを選択し、その行アドレスに対応する検出信号をアドレスイベント検出部131に出力させる。列駆動回路136は、列アドレスを選択し、その列アドレスに対応する検出信号をアドレスイベント検出部131に出力させる。信号処理回路137は、アドレスイベント検出部131からの検出信号に対し、所定の信号処理を実行する。信号処理回路137は、検出信号を画素信号として二次元格子状に配列し、画素ごとに1ビットの情報を有する画像データを取得する。信号処理回路137は、この画像データに対し、画像認識処理などの信号処理を実行する。 The row driving circuit 135 selects a row address and outputs a detection signal corresponding to that row address to the address event detection unit 131. The column driving circuit 136 selects a column address and outputs a detection signal corresponding to that column address to the address event detection unit 131. The signal processing circuit 137 performs predetermined signal processing on the detection signal from the address event detection unit 131. The signal processing circuit 137 arranges the detection signals as pixel signals in a two-dimensional lattice pattern and acquires image data having one bit of information for each pixel. The signal processing circuit 137 performs signal processing such as image recognition processing on this image data.
 図18は、第3実施形態の各アドレスイベント検出回路131aの構成を示す回路図である。 FIG. 18 is a circuit diagram showing the configuration of each address event detection circuit 131a in the third embodiment.
 各アドレスイベント検出回路131aは、電流電圧変換回路310と、バッファ320と、減算器330と、量子化器340と、転送回路350とを備えている。 Each address event detection circuit 131a includes a current-voltage conversion circuit 310, a buffer 320, a subtractor 330, a quantizer 340, and a transfer circuit 350.
 電流電圧変換回路310は、対応するフォトダイオード121aからの光電流を電圧信号に変換する。電流電圧変換回路310は、この電圧信号をバッファ320に供給する。 The current-voltage conversion circuit 310 converts the photocurrent from the corresponding photodiode 121a into a voltage signal. The current-voltage conversion circuit 310 supplies this voltage signal to the buffer 320.
 バッファ320は、電流電圧変換回路310からの電圧信号を補正する。バッファ320は、補正後の電圧信号を減算器330に出力する。 The buffer 320 corrects the voltage signal from the current-voltage conversion circuit 310. The buffer 320 outputs the corrected voltage signal to the subtractor 330.
 減算器330は、行駆動回路135からの行駆動信号に従って、バッファ320からの電圧信号のレベルを低下させる。減算器330は、低下後の電圧信号を量子化器340に供給する。 The subtractor 330 reduces the level of the voltage signal from the buffer 320 in accordance with the row drive signal from the row drive circuit 135. The subtractor 330 supplies the reduced voltage signal to the quantizer 340.
 量子化器340は、減算器330からの電圧信号をデジタル信号に量子化して検出信号として出力する。量子化器340は、この検出信号を転送回路350に出力する。 The quantizer 340 quantizes the voltage signal from the subtractor 330 into a digital signal and outputs it as a detection signal. The quantizer 340 outputs this detection signal to the transfer circuit 350.
 転送回路350は、列駆動回路136からの列駆動信号に従って、量子化器340からの検出信号を信号処理回路137に転送する。 The transfer circuit 350 transfers the detection signal from the quantizer 340 to the signal processing circuit 137 in accordance with the column drive signal from the column drive circuit 136.
 図19は、第3実施形態の電流電圧変換回路310の構成を示す回路図である。 FIG. 19 is a circuit diagram showing the configuration of a current-voltage conversion circuit 310 according to the third embodiment.
 電流電圧変換回路310は、N型トランジスタ311と、P型トランジスタ312と、N型トランジスタ313とを備えている。これらのN型およびP型トランジスタ311~313は例えば、MOS(Metal-Oxide-Semiconductor)トランジスタである。 The current-voltage conversion circuit 310 includes an N-type transistor 311, a P-type transistor 312, and an N-type transistor 313. These N-type and P-type transistors 311 to 313 are, for example, MOS (Metal-Oxide-Semiconductor) transistors.
 N型トランジスタ311のソースは、フォトダイオード121aのカソードに電気的に接続されており、N型トランジスタ311のドレインは、電源端子(VDD)に電気的に接続されている。P型トランジスタ312とN型トランジスタ313は、電源端子と接地端子(GND)との間で直列に接続されている。P型トランジスタ312とN型トランジスタ313との間のノードは、N型トランジスタ311のゲートおよびバッファ320の入力端子に電気的に接続されている。P型トランジスタ312のゲートには、所定のバイアス電圧Vbias1が印加される。N型トランジスタ311とフォトダイオード121aとの間のノードは、N型トランジスタ313のゲートと電気的に接続されている。 The source of the N-type transistor 311 is electrically connected to the cathode of the photodiode 121a, and the drain of the N-type transistor 311 is electrically connected to the power supply terminal (VDD). The P-type transistor 312 and the N-type transistor 313 are connected in series between the power supply terminal and the ground terminal (GND). The node between the P-type transistor 312 and the N-type transistor 313 is electrically connected to the gate of the N-type transistor 311 and the input terminal of the buffer 320. A predetermined bias voltage Vbias1 is applied to the gate of the P-type transistor 312. The node between the N-type transistor 311 and the photodiode 121a is electrically connected to the gate of the N-type transistor 313.
 N型トランジスタ311のドレインと、N型トランジスタ313のドレインは、電源側に配置されており、このような回路はソースフォロワと呼ばれる。フォトダイオード121aからの光電流は、ソースフォロワにより電圧信号に変換される。P型トランジスタ312は、一定の電流をN型トランジスタ313に供給する。なお、受光チップ120のグランドと検出チップ130のグランドは、干渉対策のために互いに分離されている。 The drain of N-type transistor 311 and the drain of N-type transistor 313 are placed on the power supply side, and this type of circuit is called a source follower. The photocurrent from photodiode 121a is converted into a voltage signal by the source follower. P-type transistor 312 supplies a constant current to N-type transistor 313. Note that the ground of the light receiving chip 120 and the ground of the detection chip 130 are separated from each other to prevent interference.
 図20は、第3実施形態の減算器330と量子化器340の構成を示す回路図である。 FIG. 20 is a circuit diagram showing the configuration of the subtractor 330 and quantizer 340 of the third embodiment.
 減算器330は、コンデンサ331と、インバータ332と、コンデンサ333と、スイッチ334とを備えている。量子化器340は、コンパレータ341を備えている。 The subtractor 330 includes a capacitor 331, an inverter 332, a capacitor 333, and a switch 334. The quantizer 340 includes a comparator 341.
 コンデンサ331の一方の電極は、バッファ320の出力端子に電気的に接続されており、コンデンサ331の他方の電極は、インバータ332の入力端子に電気的に接続されている。インバータ332は、コンデンサ331を介して入力された電圧信号を反転し、反転した信号をコンパレータ341の非反転入力端子(+)に出力する。コンデンサ333は、インバータ332に並列に接続されている。スイッチ334は、コンデンサ333の両電極を電気的に接続する経路を、行駆動信号に従って開閉する。 One electrode of the capacitor 331 is electrically connected to the output terminal of the buffer 320, and the other electrode of the capacitor 331 is electrically connected to the input terminal of the inverter 332. The inverter 332 inverts the voltage signal input via the capacitor 331 and outputs the inverted signal to the non-inverting input terminal (+) of the comparator 341. The capacitor 333 is connected in parallel to the inverter 332. The switch 334 opens and closes the path that electrically connects both electrodes of the capacitor 333 in accordance with the row drive signal.
 スイッチ334をオンした際に、コンデンサ331のバッファ320側の電極に電圧信号Vinitが入力され、コンデンサ331のインバータ332側の電極は仮想接地端子となる。この仮想接地端子の電位を、便宜上ゼロとする。このとき、コンデンサ331に蓄積される電荷Qinitは、コンデンサ331の容量をC1とすると、次の式1により表される。 When switch 334 is turned on, a voltage signal Vinit is input to the electrode of capacitor 331 on the buffer 320 side, and the electrode of capacitor 331 on the inverter 332 side becomes a virtual ground terminal. For convenience, the potential of this virtual ground terminal is set to zero. At this time, the charge Qinit stored in capacitor 331 is expressed by the following equation 1, where the capacitance of capacitor 331 is C1.
   Qinit=C1×Vinit ・・・ 式1
 一方、コンデンサ333の両電極は短絡されているため、コンデンサ333に蓄積される電荷はゼロとなる。
Qinit = C1 × Vinit ... Equation 1
On the other hand, since both electrodes of the capacitor 333 are short-circuited, the charge stored in the capacitor 333 becomes zero.
 次に、スイッチ334がオフされて、コンデンサ331のバッファ320側の電極の電圧が、Vafterに変化した場合を考える。このとき、コンデンサ331に蓄積される電荷Qafterは、次の式2により表される。 Next, consider the case where switch 334 is turned off and the voltage of the electrode of capacitor 331 on the buffer 320 side changes to Vafter. At this time, the charge Qafter stored in capacitor 331 is expressed by the following equation 2.
   Qafter=C1×Vafter ・・・ 式2
 一方、コンデンサ333に蓄積される電荷Q2は、出力電圧をVoutとし、コンデンサ333の容量をC2とすると、次の式3により表される。
Qafter = C1 x Vafter ... Formula 2
On the other hand, the charge Q2 stored in the capacitor 333 is expressed by the following equation 3, where the output voltage is Vout and the capacitance of the capacitor 333 is C2.
   Q2=C2×Vout ・・・ 式3
 このとき、コンデンサ331およびコンデンサ333の総電荷量は変化しないため、次の式4が成立する。
Q2 = C2 x Vout ... Equation 3
At this time, the total charge amount of the capacitors 331 and 333 does not change, so the following equation 4 holds.
   Qinit=Qafter+Q2 ・・・ 式4
 式4に式1~3を代入して変形すると、次の式5が得られる。
Qinit = Qafter + Q2 ... Equation 4
By substituting equations 1 to 3 into equation 4 and rearranging it, the following equation 5 is obtained.
   Vout=(C1/C2)×(Vinit-Vafter) ・・・ 式5
 式5は、電圧信号の減算動作を表し、減算結果の利得はC1/C2となる。通常、利得を最大化することが望まれるため、C1を大きく設定し、C2を小さく設計することが好ましい。一方、C2が小さすぎると、kTCノイズが増大し、ノイズ特性が悪化するおそれがあるため、C2の容量削減は、ノイズを許容することができる範囲に制限される。さらには、画素ごとに減算器330を含むアドレスイベント検出回路131aが搭載されるため、C1およびC2には、面積上の制約がある。これらを考慮して、例えば、C1は、20~200フェムトファラッド(fF)の値に設定され、C2は、1~20フェムトファラッド(fF)の値に設定される。
Vout = (C1/C2) x (Vinit-Vafter) ... Equation 5
Equation 5 represents the subtraction operation of the voltage signal, and the gain of the subtraction result is C1/C2. Since it is usually desired to maximize the gain, it is preferable to set C1 large and design C2 small. On the other hand, if C2 is too small, kTC noise increases and noise characteristics may deteriorate, so the reduction in the capacitance of C2 is limited to a range in which noise can be tolerated. Furthermore, since an address event detection circuit 131a including a subtractor 330 is mounted for each pixel, there are area restrictions on C1 and C2. Taking these into consideration, for example, C1 is set to a value of 20 to 200 femtofarads (fF), and C2 is set to a value of 1 to 20 femtofarads (fF).
 コンパレータ341は、減算器330からの電圧信号と、反転入力端子(-)に印加された所定の閾値電圧Vthとを比較する。コンパレータ341は、比較結果を示す信号を検出信号として転送回路350に出力する。 Comparator 341 compares the voltage signal from subtractor 330 with a predetermined threshold voltage Vth applied to the inverting input terminal (-). Comparator 341 outputs a signal indicating the comparison result as a detection signal to transfer circuit 350.
 同期型の固体撮像素子では、フォトダイオードと3個~4個のトランジスタとを含む簡易な画素回路が、画素ごとに設けられる。これに対し、非同期型の固体撮像素子100では、図18~図20に例示したように、フォトダイオード121aとアドレスイベント検出回路131aとを含む複雑な画素回路が、画素ごとに設けられる。よって、仮にフォトダイオード121aとアドレスイベント検出回路131aの両方を同じチップに配置すると、実装面積が同期型の場合よりも広くなってしまう。そこで、本実施形態の固体撮像素子100では、フォトダイオード121aおよびアドレスイベント検出回路131aをそれぞれ受光チップ120および検出チップ130に配置している。本実施形態によれば、このようなフォトダイオード121aおよびアドレスイベント検出回路131aの分散配置により、実装面積を削減することが可能となる。 In a synchronous solid-state imaging device, a simple pixel circuit including a photodiode and three to four transistors is provided for each pixel. In contrast, in an asynchronous solid-state imaging device 100, as illustrated in Figures 18 to 20, a complex pixel circuit including a photodiode 121a and an address event detection circuit 131a is provided for each pixel. Therefore, if both the photodiode 121a and the address event detection circuit 131a were placed on the same chip, the mounting area would be larger than in the case of the synchronous type. Therefore, in the solid-state imaging device 100 of this embodiment, the photodiode 121a and the address event detection circuit 131a are placed on the light receiving chip 120 and the detection chip 130, respectively. According to this embodiment, the mounting area can be reduced by distributing the photodiode 121a and the address event detection circuit 131a in this manner.
 図21は、第3実施形態の変形例の受光チップ120と検出チップ130の構成を示す回路図である。 FIG. 21 is a circuit diagram showing the configuration of the light receiving chip 120 and the detection chip 130 of a modified example of the third embodiment.
 図19では、電流電圧変換回路310内のN型トランジスタ311、P型トランジスタ312、およびN型トランジスタ313が、検出チップ130内に配置されている。一方で、図21では、電流電圧変換回路310内のN型トランジスタ311およびN型トランジスタ313が、受光チップ120内に配置され、電流電圧変換回路310内のP型トランジスタ312が、検出チップ130内に配置されている。図21の構成は例えば、画素数の増大に伴って、検出チップ130の回路規模が増大するおそれがある場合に採用される。本変形例によれば、N型トランジスタ311およびN型トランジスタ313を受光チップ120内に配置することにより、検出チップ130の回路規模を削減することが可能となる。 19, N-type transistor 311, P-type transistor 312, and N-type transistor 313 in current-voltage conversion circuit 310 are arranged in detection chip 130. On the other hand, in FIG. 21, N-type transistor 311 and N-type transistor 313 in current-voltage conversion circuit 310 are arranged in light-receiving chip 120, and P-type transistor 312 in current-voltage conversion circuit 310 is arranged in detection chip 130. The configuration in FIG. 21 is adopted, for example, when there is a risk that the circuit scale of detection chip 130 will increase with an increase in the number of pixels. According to this modified example, by arranging N-type transistor 311 and N-type transistor 313 in light-receiving chip 120, it is possible to reduce the circuit scale of detection chip 130.
 また、仮にN型トランジスタ311、313の一方を受光チップ120内に配置し、N型トランジスタ311、313の他方を検出チップ130内を配置すると、受光チップ120内にN型トランジスタ120を形成する工程と、検出チップ130内にN型トランジスタ130を形成する工程とを行う必要があり、受光チップ120および検出チップ130を製造する工程数が多くなる。本変形例によれば、N型トランジスタ311、313の両方を受光チップ120内に配置することで、受光チップ120および検出チップ130を製造する工程数を少なくすることが可能となる。これにより、固体撮像装置100の製造コストを削減することが可能となる。 Furthermore, if one of the N- type transistors 311, 313 were placed in the light receiving chip 120 and the other N- type transistor 311, 313 were placed in the detection chip 130, a process of forming the N-type transistor 120 in the light receiving chip 120 and a process of forming the N-type transistor 130 in the detection chip 130 would be required, increasing the number of processes for manufacturing the light receiving chip 120 and the detection chip 130. According to this modified example, by placing both N- type transistors 311, 313 in the light receiving chip 120, it is possible to reduce the number of processes for manufacturing the light receiving chip 120 and the detection chip 130. This makes it possible to reduce the manufacturing costs of the solid-state imaging device 100.
 以上のように、本実施形態によれば、第1または第2実施形態の固体撮像装置100の実装面積や製造コストを削減することが可能となる。 As described above, according to this embodiment, it is possible to reduce the mounting area and manufacturing costs of the solid-state imaging device 100 of the first or second embodiment.
 以上、本開示の実施形態について説明したが、これらの実施形態は、本開示の要旨を逸脱しない範囲内で、種々の変更を加えて実施してもよい。例えば、2つ以上の実施形態を組み合わせて実施してもよい。 The above describes embodiments of the present disclosure, but these embodiments may be implemented with various modifications without departing from the spirit of the present disclosure. For example, two or more embodiments may be implemented in combination.
 なお、本開示は、以下のような構成を取ることもできる。 In addition, this disclosure can also be configured as follows:
 (1)
 イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、
 前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、
 前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部と、
 を備える固体撮像装置。
(1)
a plurality of pixels that detect an event and output event data indicative of the detection result of the event;
an event generating unit that sets the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer equal to or greater than 2) from the event data of the first octave;
an event output unit that outputs at least a part of the event data of the first to i-th octaves;
A solid-state imaging device comprising:
 (2)
 前記第1~第iオクターブのイベントデータにそれぞれ、前記第1~第iオクターブの識別情報であるオクターブ情報を付加するオクターブ情報付加部をさらに備え、
 前記イベント出力部は、前記オクターブ情報が付加された前記イベントデータを出力する、(1)に記載の固体撮像装置。
(2)
an octave information adding unit that adds octave information, which is identification information of the first to i-th octaves, to the event data of the first to i-th octaves, respectively;
The solid-state imaging device according to (1), wherein the event output unit outputs the event data to which the octave information is added.
 (3)
 前記イベント出力部は、前記イベントデータをオクターブごとに画像表現で出力する、(1)に記載の固体撮像装置。
(3)
The solid-state imaging device according to (1), wherein the event output unit outputs the event data in an image representation for each octave.
 (4)
 前記イベント生成部は、第jオクターブ(jは1≦j≦i-1を満たす整数)のイベントデータから第j+1オクターブのイベントデータを生成する、(1)に記載の固体撮像装置。
(4)
The solid-state imaging device according to (1), wherein the event generating unit generates event data of a j+1-th octave from event data of a j-th octave (j is an integer satisfying 1≦j≦i−1).
 (5)
 前記イベント生成部は、m行分(mは2以上の整数)の前記第jオクターブのイベントデータから、1行分の前記第j+1オクターブのイベントデータを生成する、(4)に記載の固体撮像装置。
(5)
The solid-state imaging device according to (4), wherein the event generation unit generates one row of event data for the j+1 octave from m rows of event data for the j octave (m is an integer equal to or greater than 2).
 (6)
 前記イベント生成部は、m行n列分(nは2以上の整数)の前記第jオクターブのイベントデータから、1行1列分の前記第j+1オクターブのイベントデータを生成する、(5)に記載の固体撮像装置。
(6)
The solid-state imaging device according to (5), wherein the event generation unit generates event data for the j+1 octave for 1 row and 1 column from event data for the j octave for m rows and n columns (n is an integer equal to or greater than 2).
 (7)
 前記イベント生成部は、前記m行n列分の前記第jオクターブのイベントデータがk個のイベント発火(kは1≦k≦m×nを満たす整数)を含む場合に、前記1行1列分の前記第j+1オクターブのイベントデータにてイベントを発火させる、(6)に記載の固体撮像装置。
(7)
The solid-state imaging device according to (6), wherein the event generation unit ignites an event with the event data of the j+1 octave of the 1st row and 1st column when the event data of the j octave of the m rows and n columns includes k event firings (k is an integer satisfying 1≦k≦m×n).
 (8)
 mは2であり、nは2であり、かつkは1、2、3、および4である、(7)に記載の固体撮像装置。
(8)
The solid-state imaging device according to (7), wherein m is 2, n is 2, and k is 1, 2, 3, or 4.
 (9)
 mは2であり、nは2であり、かつkは2、3、および4である、(7)に記載の固体撮像装置。
(9)
The solid-state imaging device according to (7), wherein m is 2, n is 2, and k is 2, 3, or 4.
 (10)
 mは2であり、nは2であり、かつkは3および4である、(7)に記載の固体撮像装置。
(10)
The solid-state imaging device according to (7), wherein m is 2, n is 2, and k is 3 or 4.
 (11)
 前記複数の画素から出力された前記イベントデータを格納するフレームメモリをさらに備え、
 前記イベント生成部は、前記フレームメモリから出力された前記イベントデータを前記第1オクターブのイベントデータとする、(1)に記載の固体撮像装置。
(11)
A frame memory for storing the event data output from the plurality of pixels,
The solid-state imaging device according to (1), wherein the event generating unit generates the event data output from the frame memory as event data of the first octave.
 (12)
 固体撮像装置と情報処理部とを備える情報処理システムであって、
 前記固体撮像装置は、
 イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、
 前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、
 前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部とを備え、
 前記情報処理部は、前記イベント出力部から出力された前記イベントデータを表示画面に表示する、情報処理システム。
(12)
An information processing system including a solid-state imaging device and an information processing unit,
The solid-state imaging device includes:
a plurality of pixels that detect an event and output event data indicative of the detection result of the event;
an event generating unit that sets the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer equal to or greater than 2) from the event data of the first octave;
an event output unit that outputs at least a part of the event data of the first to i-th octaves;
The information processing unit displays the event data output from the event output unit on a display screen.
 (13)
 前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、
 前記情報処理部は、前記抽出部により抽出された前記イベントデータを前記表示画面に表示する、(12)に記載の情報処理システム。
(13)
the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit,
The information processing system according to (12), wherein the information processing unit displays the event data extracted by the extraction unit on the display screen.
 (14)
 前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、
 前記情報処理部は、前記抽出部により抽出された前記イベントデータを利用して画像認識を行う、(12)に記載の情報処理システム。
(14)
the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit,
The information processing system according to (12), wherein the information processing unit performs image recognition using the event data extracted by the extraction unit.
 (15)
 前記画像認識は、ユーザーのジェスチャー認識である、(14)に記載の情報処理システム。
(15)
The information processing system according to (14), wherein the image recognition is user gesture recognition.
 (16)
 前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、ユーザーにより指定されたオクターブ数の前記イベントデータを選択する選択部を含み、
 前記情報処理部は、前記選択部により選択された前記イベントデータを記録媒体に記録する、(12)に記載の情報処理システム。
(16)
the information processing unit includes a selection unit that selects the event data for a number of octaves designated by a user from the event data output from the event output unit,
The information processing system according to (12), wherein the information processing unit records the event data selected by the selection unit on a recording medium.
 (17)
 前記情報処理システムは、前記固体撮像装置と前記情報処理部とを備える電子機器である、(12)に記載の情報処理システム。
(17)
The information processing system according to (12), wherein the information processing system is an electronic device including the solid-state imaging device and the information processing unit.
 (18)
 前記電子機器はさらに、前記表示画面を有する表示部を備える、(17)に記載の情報処理システム。
(18)
The information processing system according to (17), wherein the electronic device further includes a display unit having the display screen.
 (19)
 前記情報処理システムは、前記情報処理部を含む電子機器と、前記電子機器の外部に設けられ、前記固体撮像装置を含む撮像装置とを備える、(12)に記載の情報処理システム。
(19)
The information processing system according to (12), further comprising: an electronic device including the information processing section; and an imaging device provided outside the electronic device and including the solid-state imaging device.
 (20)
 前記情報処理システムはさらに、前記電子機器の外部に設けられ、前記表示画面を有する表示装置を備える、(19)に記載の情報処理システム。
(20)
The information processing system according to (19), further comprising a display device provided outside the electronic device and having the display screen.
 1:車両、11:車両制御システム、21:車両制御ECU、
 22:通信部、23:地図情報蓄積部、24:位置情報取得部、
 25:外部認識センサ、26:車内センサ、27:車両センサ、
 31:記憶部、32:走行支援・自動運転制御部、33:DMS、
 34:HMI、35:車両制御部、41:通信ネットワーク、
 51:カメラ、52:レーダ、53:LiDAR、54:超音波センサ、
 61:分析部、62:行動計画部、63:動作制御部、
 71:自己位置推定部、72:センサフュージョン部、73:認識部、
 81:ステアリング制御部、82:ブレーキ制御部、83:駆動制御部、
 84:ボディ系制御部、85:ライト制御部、86:ホーン制御部、
 100:固体撮像装置、101:画素アレイ、101a:画素、
 102:イベント取得部、102a:アービタ部、102b:タイムスタンプ部、
 103:イベント生成部、103a:第1フィルタ部、103b:第2フィルタ部、
 103c:第3フィルタ部、104:イベント合成部、
 104a:オクターブ情報付加部、104b:出力タイミング調整部、
 105:イベント出力部、105a:イベントデータ選択部、
 105b:イベントデータ形成部、111:フレームメモリ、
 120:受光チップ、121:受光部、121a:フォトダイオード、
 122:ビア配置部、123:ビア配置部、124:ビア配置部、
 130:検出チップ、131:アドレスイベント検出部、
 131a:アドレスイベント検出回路、132:ビア配置部、
 133:ビア配置部、134:ビア配置部、
 135:行駆動回路、136:列駆動回路、137:信号処理回路、
 200:電子機器、201:撮像部、201’:撮像装置、
 202:表示部、202’:表示装置、203:情報処理部、
 203a:抽出部、203b:選択部、204:記憶部、205:入力部、
 310:電流電圧変換回路、311:N型トランジスタ、
 312:P型トランジスタ、313:N型トランジスタ、
 320:バッファ、330:減算器、331:コンデンサ、
 332:インバータ、333:コンデンサ、334:スイッチ、
 340:量子化器、341:コンパレータ、350:転送回路
1: vehicle, 11: vehicle control system, 21: vehicle control ECU,
22: communication unit, 23: map information storage unit, 24: location information acquisition unit,
25: external recognition sensor, 26: in-vehicle sensor, 27: vehicle sensor,
31: memory unit, 32: driving assistance/automatic driving control unit, 33: DMS,
34: HMI, 35: vehicle control unit, 41: communication network,
51: camera, 52: radar, 53: LiDAR, 54: ultrasonic sensor,
61: analysis unit, 62: action planning unit, 63: operation control unit,
71: self-position estimation unit, 72: sensor fusion unit, 73: recognition unit,
81: steering control unit, 82: brake control unit, 83: drive control unit,
84: body control unit, 85: light control unit, 86: horn control unit,
100: solid-state imaging device, 101: pixel array, 101a: pixel,
102: event acquisition unit, 102a: arbiter unit, 102b: time stamp unit,
103: event generating unit, 103a: first filter unit, 103b: second filter unit,
103c: third filter unit, 104: event synthesis unit,
104a: octave information adding unit, 104b: output timing adjusting unit,
105: event output unit, 105a: event data selection unit,
105b: event data forming unit, 111: frame memory,
120: light receiving chip, 121: light receiving unit, 121a: photodiode,
122: via placement unit, 123: via placement unit, 124: via placement unit,
130: detection chip, 131: address event detection unit,
131a: address event detection circuit, 132: via placement unit,
133: via arrangement section, 134: via arrangement section,
135: row driving circuit, 136: column driving circuit, 137: signal processing circuit,
200: electronic device, 201: imaging unit, 201': imaging device,
202: display unit, 202': display device, 203: information processing unit,
203a: extraction unit, 203b: selection unit, 204: storage unit, 205: input unit,
310: current-voltage conversion circuit, 311: N-type transistor,
312: P-type transistor, 313: N-type transistor,
320: buffer, 330: subtractor, 331: capacitor,
332: inverter, 333: capacitor, 334: switch,
340: quantizer, 341: comparator, 350: transfer circuit

Claims (20)

  1.  イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、
     前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、
     前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部と、
     を備える固体撮像装置。
    a plurality of pixels that detect an event and output event data indicative of the detection result of the event;
    an event generating unit that sets the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer equal to or greater than 2) from the event data of the first octave;
    an event output unit that outputs at least a part of the event data of the first to i-th octaves;
    A solid-state imaging device comprising:
  2.  前記第1~第iオクターブのイベントデータにそれぞれ、前記第1~第iオクターブの識別情報であるオクターブ情報を付加するオクターブ情報付加部をさらに備え、
     前記イベント出力部は、前記オクターブ情報が付加された前記イベントデータを出力する、請求項1に記載の固体撮像装置。
    an octave information adding unit that adds octave information, which is identification information of the first to i-th octaves, to the event data of the first to i-th octaves, respectively;
    The solid-state imaging device according to claim 1 , wherein the event output section outputs the event data to which the octave information is added.
  3.  前記イベント出力部は、前記イベントデータをオクターブごとに画像表現で出力する、請求項1に記載の固体撮像装置。 The solid-state imaging device according to claim 1, wherein the event output unit outputs the event data in an image representation for each octave.
  4.  前記イベント生成部は、第jオクターブ(jは1≦j≦i-1を満たす整数)のイベントデータから第j+1オクターブのイベントデータを生成する、請求項1に記載の固体撮像装置。 The solid-state imaging device according to claim 1, wherein the event generating unit generates event data for the j+1th octave from event data for the jth octave (j is an integer satisfying 1≦j≦i-1).
  5.  前記イベント生成部は、m行分(mは2以上の整数)の前記第jオクターブのイベントデータから、1行分の前記第j+1オクターブのイベントデータを生成する、請求項4に記載の固体撮像装置。 The solid-state imaging device according to claim 4, wherein the event generation unit generates one row of event data for the j+1 octave from m rows of event data for the j octave (m is an integer equal to or greater than 2).
  6.  前記イベント生成部は、m行n列分(nは2以上の整数)の前記第jオクターブのイベントデータから、1行1列分の前記第j+1オクターブのイベントデータを生成する、請求項5に記載の固体撮像装置。 The solid-state imaging device according to claim 5, wherein the event generation unit generates 1 row and 1 column of event data for the j+1 octave from m rows and n columns of event data for the j octave (n is an integer equal to or greater than 2).
  7.  前記イベント生成部は、前記m行n列分の前記第jオクターブのイベントデータがk個のイベント発火(kは1≦k≦m×nを満たす整数)を含む場合に、前記1行1列分の前記第j+1オクターブのイベントデータにてイベントを発火させる、請求項6に記載の固体撮像装置。 The solid-state imaging device according to claim 6, wherein the event generating unit ignites an event with the event data of the j+1 octave of the 1st row and 1st column when the event data of the j octave of the m rows and n columns includes k event firings (k is an integer satisfying 1≦k≦m×n).
  8.  mは2であり、nは2であり、かつkは1、2、3、および4である、請求項7に記載の固体撮像装置。 The solid-state imaging device of claim 7, wherein m is 2, n is 2, and k is 1, 2, 3, and 4.
  9.  mは2であり、nは2であり、かつkは2、3、および4である、請求項7に記載の固体撮像装置。 The solid-state imaging device of claim 7, wherein m is 2, n is 2, and k is 2, 3, and 4.
  10.  mは2であり、nは2であり、かつkは3および4である、請求項7に記載の固体撮像装置。 The solid-state imaging device of claim 7, wherein m is 2, n is 2, and k is 3 and 4.
  11.  前記複数の画素から出力された前記イベントデータを格納するフレームメモリをさらに備え、
     前記イベント生成部は、前記フレームメモリから出力された前記イベントデータを前記第1オクターブのイベントデータとする、請求項1に記載の固体撮像装置。
    A frame memory for storing the event data output from the plurality of pixels,
    2. The solid-state imaging device according to claim 1, wherein the event generating section generates the event data output from the frame memory as the event data of the first octave.
  12.  固体撮像装置と情報処理部とを備える情報処理システムであって、
     前記固体撮像装置は、
     イベントを検出し、前記イベントの検出結果を示すイベントデータを出力する複数の画素と、
     前記複数の画素から出力された前記イベントデータを第1オクターブのイベントデータとし、前記第1オクターブのイベントデータから第2~第iオクターブ(iは2以上の整数)のイベントデータを生成するイベント生成部と、
     前記第1~第iオクターブのイベントデータのうちの少なくとも一部のイベントデータを出力するイベント出力部とを備え、
     前記情報処理部は、前記イベント出力部から出力された前記イベントデータを表示画面に表示する、情報処理システム。
    An information processing system including a solid-state imaging device and an information processing unit,
    The solid-state imaging device includes:
    a plurality of pixels that detect an event and output event data indicative of the detection result of the event;
    an event generating unit that sets the event data output from the plurality of pixels as event data of a first octave and generates event data of a second to i-th octave (i is an integer equal to or greater than 2) from the event data of the first octave;
    an event output unit that outputs at least a part of the event data of the first to i-th octaves;
    The information processing unit displays the event data output from the event output unit on a display screen.
  13.  前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、
     前記情報処理部は、前記抽出部により抽出された前記イベントデータを前記表示画面に表示する、請求項12に記載の情報処理システム。
    the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit,
    The information processing system according to claim 12 , wherein the information processing unit displays the event data extracted by the extraction unit on the display screen.
  14.  前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、所定のオクターブ数の前記イベントデータを抽出する抽出部を含み、
     前記情報処理部は、前記抽出部により抽出された前記イベントデータを利用して画像認識を行う、請求項12に記載の情報処理システム。
    the information processing unit includes an extraction unit that extracts the event data of a predetermined number of octaves from the event data output from the event output unit,
    The information processing system according to claim 12 , wherein the information processing unit performs image recognition using the event data extracted by the extraction unit.
  15.  前記画像認識は、ユーザーのジェスチャー認識である、請求項14に記載の情報処理システム。 The information processing system according to claim 14, wherein the image recognition is user gesture recognition.
  16.  前記情報処理部は、前記イベント出力部から出力された前記イベントデータから、ユーザーにより指定されたオクターブ数の前記イベントデータを選択する選択部を含み、
     前記情報処理部は、前記選択部により選択された前記イベントデータを記録媒体に記録する、請求項12に記載の情報処理システム。
    the information processing unit includes a selection unit that selects the event data for a number of octaves designated by a user from the event data output from the event output unit,
    The information processing system according to claim 12 , wherein the information processing unit records the event data selected by the selection unit on a recording medium.
  17.  前記情報処理システムは、前記固体撮像装置と前記情報処理部とを備える電子機器である、請求項12に記載の情報処理システム。 The information processing system according to claim 12, wherein the information processing system is an electronic device including the solid-state imaging device and the information processing unit.
  18.  前記電子機器はさらに、前記表示画面を有する表示部を備える、請求項17に記載の情報処理システム。 The information processing system according to claim 17, wherein the electronic device further comprises a display unit having the display screen.
  19.  前記情報処理システムは、前記情報処理部を含む電子機器と、前記電子機器の外部に設けられ、前記固体撮像装置を含む撮像装置とを備える、請求項12に記載の情報処理システム。 The information processing system according to claim 12, comprising an electronic device including the information processing section, and an imaging device provided outside the electronic device and including the solid-state imaging device.
  20.  前記情報処理システムはさらに、前記電子機器の外部に設けられ、前記表示画面を有する表示装置を備える、請求項19に記載の情報処理システム。 The information processing system according to claim 19, further comprising a display device provided outside the electronic device and having the display screen.
PCT/JP2023/037961 2022-11-14 2023-10-20 Solid-state imaging device and information processing system WO2024106132A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-182108 2022-11-14
JP2022182108A JP2024071260A (en) 2022-11-14 2022-11-14 Solid-state imaging device and information processing system

Publications (1)

Publication Number Publication Date
WO2024106132A1 true WO2024106132A1 (en) 2024-05-23

Family

ID=91084220

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/037961 WO2024106132A1 (en) 2022-11-14 2023-10-20 Solid-state imaging device and information processing system

Country Status (2)

Country Link
JP (1) JP2024071260A (en)
WO (1) WO2024106132A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017533497A (en) * 2014-09-16 2017-11-09 クゥアルコム・インコーポレイテッドQualcomm Incorporated Event-based downsampling
JP2020136958A (en) * 2019-02-21 2020-08-31 ソニーセミコンダクタソリューションズ株式会社 Event signal detection sensor and control method
JP2020161992A (en) * 2019-03-27 2020-10-01 ソニーセミコンダクタソリューションズ株式会社 Imaging system and object recognition system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017533497A (en) * 2014-09-16 2017-11-09 クゥアルコム・インコーポレイテッドQualcomm Incorporated Event-based downsampling
JP2020136958A (en) * 2019-02-21 2020-08-31 ソニーセミコンダクタソリューションズ株式会社 Event signal detection sensor and control method
JP2020161992A (en) * 2019-03-27 2020-10-01 ソニーセミコンダクタソリューションズ株式会社 Imaging system and object recognition system

Also Published As

Publication number Publication date
JP2024071260A (en) 2024-05-24

Similar Documents

Publication Publication Date Title
JP7314798B2 (en) IMAGING DEVICE, IMAGE PROCESSING DEVICE, AND IMAGE PROCESSING METHOD
WO2020195934A1 (en) Data processing device, data processing method, and program
US11815799B2 (en) Information processing apparatus and information processing method, imaging apparatus, mobile device, and computer program
WO2020195700A1 (en) Image capturing device, and image capturing system
WO2021241189A1 (en) Information processing device, information processing method, and program
WO2022075133A1 (en) Imaging device, information processing device, imaging system, and imaging method
WO2024024148A1 (en) On-vehicle monitoring device, information processing device, and on-vehicle monitoring system
WO2020116205A1 (en) Information processing device, information processing method, and program
WO2022153896A1 (en) Imaging device, image processing method, and image processing program
WO2024106132A1 (en) Solid-state imaging device and information processing system
WO2022004423A1 (en) Information processing device, information processing method, and program
WO2020158489A1 (en) Visible light communication device, visible light communication method, and visible light communication program
WO2024062842A1 (en) Solid-state imaging device
WO2024106196A1 (en) Solid-state imaging device and electronic apparatus
US20210295563A1 (en) Image processing apparatus, image processing method, and program
US20240179429A1 (en) Solid-state imaging device, imaging device, processing method in solid-state imaging device, processing program in solid-state imaging device, processing method in imaging device, and processing program in imaging device
WO2022085479A1 (en) Information processing device, information processing method, and program
WO2022153888A1 (en) Solid-state imaging device, control method for solid-state imaging device, and control program for solid-state imaging device
WO2023054090A1 (en) Recognition processing device, recognition processing method, and recognition processing system
WO2024009739A1 (en) Optical ranging sensor and optical ranging system
WO2022075039A1 (en) Information processing device, information processing system, and information processing method
WO2024062976A1 (en) Information processing device and information processing method
WO2024018920A1 (en) Information processing device and information processing method
WO2023189071A1 (en) Imaging device and electronic apparatus
WO2024024471A1 (en) Information processing device, information processing method, and information processing system