US20190235823A1 - Projection device and control method therefor - Google Patents
Projection device and control method therefor Download PDFInfo
- Publication number
- US20190235823A1 US20190235823A1 US16/377,914 US201916377914A US2019235823A1 US 20190235823 A1 US20190235823 A1 US 20190235823A1 US 201916377914 A US201916377914 A US 201916377914A US 2019235823 A1 US2019235823 A1 US 2019235823A1
- Authority
- US
- United States
- Prior art keywords
- image
- display apparatus
- cpu
- information
- projector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1423—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
- G06F3/1446—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display display composed of modules, e.g. video walls
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/64—Constructional details of receivers, e.g. cabinets or dust covers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3141—Constructional details thereof
- H04N9/3147—Multi-projection systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3179—Video signal processing therefor
- H04N9/3185—Geometric adjustment, e.g. keystone or convergence
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/0693—Calibration of display systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2356/00—Detection of the display position w.r.t. other display screens
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/20—Details of the management of multiple sources of image data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
Definitions
- the present disclosure relates to a projection device and a control method therefor.
- a multiple projection system which displays one projection image by synthesizing, on a screen, a plurality of images projected onto the screen by each of a plurality of projectors.
- one projector (a master) of a plurality of projectors controls the other projectors (slaves).
- Japanese Patent Application Laid-Open No. 2006-338044 discloses that, in the multiple projection system, one projector (a master) of a plurality of projectors associates identification information of the other projectors (slaves) with positional relationship of the other projectors (slaves). Further, it is known that the projector (master) notifies a position in one projection image displayed on a screen by the multiple projection system, to which a position of an image projected by another projector (slave) corresponds, to another projector (slave).
- At least one aspect of the present disclosure is to solve all or at least one of the issues described above.
- a projection device includes a projection unit configured to project an image, a communication unit configured to communicate with an external projection device, and an input unit configured to input an association between the external projection device and a predetermined area included in a display pattern for displaying one image on a screen by using an image projected by the projection device and an image projected by the external projection device.
- the projection unit projects an image for notifying of the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit
- the communication unit transmits information related to the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, to the external projection device.
- FIG. 1 is a diagram illustrating an example of a multiple projection system in first to third embodiments.
- FIGS. 2A and 2B are diagrams illustrating an example of a configuration of a projector in the first to the third embodiments.
- FIG. 3 is a sequence diagram illustrating an example of projection processing performed in the first embodiment.
- FIG. 4 is a diagram illustrating an example of a selection image in the first to the third embodiments.
- FIGS. 5A and 5B are diagrams illustrating an example of an input image in the first to the third embodiments.
- FIG. 6 is a diagram illustrating an example of an input image in the first to the third embodiments.
- FIG. 7 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
- FIGS. 8A and 8B are diagrams illustrating an example of an input image in the first to the third embodiments.
- FIG. 9 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
- FIG. 10 is a sequence diagram illustrating an example of projection processing in the second embodiment.
- FIG. 11 is a diagram illustrating an example of an image projected by the multiple projection system in the second embodiment.
- FIG. 12 is a flowchart illustrating an example of change processing in the third embodiment.
- FIGS. 13A and 13B are diagrams illustrating an example of an image projected by the multiple projection system in the third embodiment.
- FIG. 1 is a diagram illustrating a multiple projection system in the first embodiment.
- the multiple projection system is a system that displays one image on a screen by synthesizing, on the screen, images projected onto the screen by each of a plurality of projectors.
- the multiple projection system of the first embodiment includes four projectors 100 a to 100 d .
- the projectors 100 a to 100 d respectively project four partial images 210 a to 210 d , which are obtained by dividing one image 200 to be displayed on a screen into four images, onto the screen.
- the partial images 210 a to 210 d are displayed on the screen and one image corresponding to the image 200 is displayed on the screen by combining the four partial images 210 a to 210 d .
- the partial image 210 a may include at least one of a part of the partial image 210 b , a part of the partial image 210 c , and a part of the partial image 210 d .
- the partial image 210 b may include at least one of a part of the partial image 210 a , a part of the partial image 210 c , and a part of the partial image 210 d .
- the partial image 210 c may include at least one of a part of the partial image 210 a , a part of the partial image 210 b , and a part of the partial image 210 d .
- the partial image 210 d may include at least one of a part of the partial image 210 a , a part of the partial image 210 b , and a part of the partial image 210 c.
- each of the projectors 100 a to 100 d projects a partial image obtained by dividing one image
- an image projected onto the screen by each of the projectors 100 a to 100 d may be an image other than a partial image obtained by dividing one image.
- the projector in the first embodiment may be rephrased to a “projection device” or may be rephrased to a “display device”.
- the number of projectors included in the multiple projection system is not limited to four, but may be two.
- the number of projectors included in the multiple projection system may be two or more.
- the projector 100 a of the projectors 100 a to 100 d is set as a master device and the other projectors 100 b to 100 d are set as slave devices.
- the projector 100 a transmits the image 200 to the projectors 100 b to 100 d .
- each of the projectors 100 b to 100 d generates a partial image from the image 200 and projects the generated partial image onto the screen.
- the projector 100 b generates a partial image 210 b from the image 200 received from the projector 100 a and projects the partial image 210 b onto the screen.
- the projector 100 c generates a partial image 210 c from the image 200 received from the projector 100 a and projects the partial image 210 c onto the screen.
- the projector 100 d generates a partial image 210 d from the image 200 received from the projector 100 a and projects the partial image 210 d onto the screen.
- FIGS. 2A and 2B are diagrams illustrating an example of a configuration of the projector 100 a .
- FIG. 2A is a diagram illustrating an example of a configuration of the projector 100 a and
- FIG. 2B is an external view illustrating an example of the projector 100 a.
- the projector 100 a includes a CPU (Central Processing Unit) 101 a , a ROM 102 a , a RAM 103 a , an image processing unit 104 a , a network IF (Interface) 105 a , and a user IF 106 a . Further, the projector 100 a includes an operation key 107 a , a remote control light-receiving unit 108 a , a projection unit 111 a , an input unit 112 a , and an internal bus 113 a.
- the CPU 101 a is a control unit that controls each unit of the projector 100 a .
- the ROM 102 a stores a computer program to be executed by the CPU 101 a .
- the RAM 103 a temporarily stores a computer program and data as a work memory.
- the CPU 101 a controls the projector 100 a by executing the program stored in the ROM 102 a.
- the image processing unit 104 a performs image processing to generate an image to be projected by the projection unit 111 a .
- the image processing unit 104 a performs light reduction processing (gradation processing) on an image that overlaps a part of an image projected by a projector other than the projector 100 a in an image projected by the projection unit 111 a .
- the image that overlaps a part of an image projected by a projector other than the projector 100 a is referred to as an “overlapping area”.
- the network IF 105 a communicates with an external projector (an external projection device) through a LAN (Local Area Network).
- the network IF 105 a may communicate with an external projector through a wireless LAN or may communicate with an external projector through a wired LAN.
- the network IF 105 a may communicate with an external projector based on a communication protocol other than LAN.
- the user IF 106 a receives an input by a user from at least either one of the operation key 107 a and a remote control unit 109 a .
- the remote control light-receiving unit 108 a supplies a signal received from the remote control unit 109 a to the user IF 106 a.
- the projection unit 111 a projects an image supplied from the image processing unit 104 a .
- the projection unit 111 a includes a liquid crystal panel and a driver that drives the liquid crystal panel. Further, the projection unit 111 a includes a drive system such as a lens and a zoom and a control system that controls a light source and a position of an image to be projected.
- the input unit 112 a receives an image from outside.
- the internal bus 113 a connects units included in the projector 100 a .
- the remote control unit 109 a is used to control the projector 100 a.
- the projectors 100 b to 100 d also have the same configuration as that of the projector 100 a .
- the projector 100 b includes a CPU 101 b , a ROM 102 b , a RAM 103 b , an image processing unit 104 b , a network IF 105 b , and a user IF 106 b .
- the projector 100 b includes an operation key 107 b , a remote control light-receiving unit 108 b , a projection unit 111 b , an input unit 112 b , and an internal bus 113 b .
- the projector 100 c includes a CPU 101 c , a ROM 102 c , a RAM 103 c , an image processing unit 104 c , a network IF 105 c , and a user IF 106 c .
- the projector 100 c includes an operation key 107 c , a remote control light-receiving unit 108 c , a projection unit 111 c , an input unit 112 c , and an internal bus 113 c .
- the projector 100 d includes a CPU 101 d , a ROM 102 d , a RAM 103 d , an image processing unit 104 d , a network IF 105 d , and a user IF 106 d .
- the projector 100 d includes an operation key 107 d , a remote control light-receiving unit 108 d , a projection unit 111 d , an input unit 112 d , and an internal bus 113 d.
- the projector 100 a supplies an image received through at least either one of the input unit 112 a and the network IF 105 a to the image processing unit 104 a . Thereafter, the projector 100 a controls the projection unit 111 a so as to project an image on which image processing is performed by the image processing unit 104 a onto a screen. While the operation of the projector 100 a has been described, the projectors 100 b to 100 d also perform the same operation as that performed by the projector 100 a.
- FIG. 3 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the first embodiment.
- the projection processing illustrated in FIG. 3 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices.
- the CPU 101 a When a user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the projector 100 a is selected as the master device, the CPU 101 a performs S 301 .
- the CPU 101 a receives a master selection instruction through the user IF 106 a .
- the CPU 101 a sets an operation mode of the projector 100 a to a master mode in which the projector 100 a operates as the master device. In this case, the CPU 101 a stores a value indicating that the projector 100 a is set to the master device into the RAM 103 a .
- the CPU 101 a After S 301 is performed, the CPU 101 a performs S 302 .
- the CPU 101 b When the user operates at least either one of the operation key 107 b and the remote control unit 109 b in the projector 100 b and thereby the projector 100 b is selected as the slave device, the CPU 101 b performs S 321 .
- the CPU 101 b receives a slave selection instruction through the user IF 106 b .
- the CPU 101 b sets an operation mode of the projector 100 b to a slave mode in which the projector 100 b operates as the slave device.
- the CPU 101 b stores a value indicating that the projector 100 b is set to the slave device into the RAM 103 b .
- the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 321 .
- the CPU 101 a controls the projection unit 111 a so as to project a selection image 400 . Thereafter, the projection unit 111 a projects the selection image 400 onto a screen.
- the selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system.
- FIG. 4 is a diagram illustrating an example of the selection image 400 .
- the selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system from among display patterns 401 to 404 .
- the display patterns 401 to 404 are images which respectively represent display patterns different from each other.
- the display pattern 401 is an image representing a pattern in which two partial images are arranged in the horizontal direction and displayed on the screen by two projectors. Further, in the display pattern 401 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the right partial image.
- the display pattern 402 is an image representing a pattern of 2 ⁇ 2.
- the display pattern 402 is an image representing a pattern in which four partial images are displayed on the screen by four projectors. Further, in the display pattern 402 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the upper left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the upper right partial image. Further, for example, identification information “3” is information for identifying a position of a predetermined area of the lower left partial image. Further, for example, identification information “4” is information for identifying a position of a predetermined area of the lower right partial image.
- the display pattern 403 is an image representing a display pattern of 2 ⁇ 3.
- the display pattern 403 is an image representing a pattern in which six partial images are displayed on the screen. Further, in the display pattern 403 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
- the display pattern 404 is an image representing a display pattern of 3 ⁇ 3.
- the display pattern 404 is an image representing a pattern in which nine partial images projected on the screen by nine projectors are displayed on the screen. Further, in the display pattern 403 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
- the display patterns 401 to 404 and the selection image 400 are stored in advance in the ROM 102 a of the projector 100 a which is the master device.
- the selection image 400 is displayed on the screen by the projection unit 111 a .
- a user selects any one of the display patterns 401 to 404 by using at least either one of the operation key 107 a and the remote control unit 109 a .
- the CPU 101 a performs S 303 .
- the CPU 101 a receives a selection instruction of a display pattern through the user IF 106 a . Thereafter, the CPU 101 a selects any one of the display patterns 401 to 404 according to the selection instruction from the user and records information indicating the selected display pattern in the RAM 103 a . After S 303 is performed, the CPU 101 a performs S 304 .
- the CPU 101 a selects the display pattern 402 and stores information indicating the display pattern 402 in the RAM 103 a.
- the CPU 101 a controls the projection unit 111 a so as to project an input image 500 . Further, the CPU 101 a secures an area to assign IP addresses to a plurality of pieces of identification information of the display pattern selected in S 302 in the RAM 103 a.
- the projection unit 111 a projects the input image 500 illustrated in FIG. 5A onto the screen in S 304 .
- the input image 500 illustrated in FIG. 5A will be described.
- the input image 500 illustrated in FIG. 5A includes an area 510 , an area 520 , and a confirmation button 530 .
- the area 510 is an area where the display pattern 402 selected in S 302 is displayed.
- the area 520 includes an area 521 where identification information for identifying positions of predetermined areas 510 a to 510 d included in the display pattern 402 is displayed and an input field 522 .
- the display pattern 402 displayed in the area 510 includes the predetermined areas 510 a to 510 d , so that pieces of identification information “1” to “4” for identifying the positions of the predetermined areas 510 a to 510 d are displayed in the area 521 .
- the input field 522 is an input field where IP addresses of the projectors are inputted with respect to the identification information “1” to “4” displayed in the area 521 .
- the IP address is identification information for identifying each of the projectors 100 a to 100 d in a network for implementing the multiple projection system.
- the area 521 includes areas 521 a to 521 d .
- the input field 522 includes input fields 522 a to 522 d .
- the identification information “1” for identifying the position of the predetermined area 510 a is displayed.
- the input field 522 a is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 a .
- the IP address inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”.
- the input field 522 b is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 b .
- the IP address inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”.
- the input field 522 c is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 c .
- the IP address inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”.
- the input field 522 d is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 d .
- the IP address inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”.
- the input image 500 is used to cause the user to input IP addresses into the input fields 522 a to 522 d.
- the input image 500 illustrated in FIG. 5A is displayed on the screen by the projection unit 111 a .
- the user inputs IP addresses into the input field 522 as illustrated in FIG. 5B by using at least either one of the operation key 107 a and the remote control unit 109 a .
- the IP address of the projector 100 a is “192.168.254.1” and the IP address of the projector 100 b is “192.168.254.2”.
- the IP address of the projector 100 c is “192.168.254.3” and the IP address of the projector 100 d is “192.168.254.4”.
- the user can input the IP addresses of the projectors into the input field 522 while confirming the positions where the projectors 100 a to 100 d are installed, the area 510 , and the area 521 .
- the IP address of the projector 100 a which is the master device, is inputted into the input field 522 a in advance even when the user does not input the IP address.
- the user should input the IP address of the projector 100 b into the input field 522 b . Further, the user should input the IP address of the projector 100 c into the input field 522 c . Further, the user should input the IP address of the projector 100 d into the input field 522 d.
- the CPU 101 a monitors a subnet of the IP addresses inputted into the input field 522 .
- the CPU 101 a controls the projection unit 111 a so as to project information indicating that there is an error in the input of the IP addresses onto the screen.
- the information 600 such as “Subnets are different” as illustrated in FIG. 6 is projected by the projection unit 111 a .
- the CPU 101 a performs S 305 .
- the CPU 101 a sets projection positions of the projectors 100 a to 100 d according to the input into the input field 522 performed by the user.
- the IP address “192.168.254.1” inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”.
- the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.1” projects a partial image, to the position indicated by the identification information “1”. Therefore, the CPU 101 a stores first position information including the identification information “1” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.1”.
- the first position information may further include information indicating the area 510 a.
- the IP address “192.168.254.2” inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”.
- the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.2” projects a partial image, to the position indicated by the identification information “2”. Therefore, the CPU 101 a stores second position information including the identification information “2” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.2”.
- the second position information may further include information indicating the predetermined area 510 b.
- the IP address “192.168.254.3” inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”.
- the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.3” projects a partial image, to the position indicated by the identification information “3”. Therefore, the CPU 101 a stores third position information including the identification information “3” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.3”.
- the third position information may further include information indicating the predetermined area 510 c.
- the IP address “192.168.254.4” inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”.
- the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.4” projects a partial image, to the position indicated by the identification information “4”. Therefore, the CPU 101 a stores fourth position information including the identification information “4” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.4”.
- the fourth position information may further include information indicating the predetermined area 510 d.
- the CPU 101 a When the user presses the confirmation button 530 by using at least either one of the operation key 107 a and the remote control unit 109 a , the CPU 101 a performs S 306 .
- the confirmation button 530 When the confirmation button 530 is pressed, the assignment of the IP addresses by using the input image 500 is completed.
- the CPU 101 a transmits the position information recorded in the RAM 103 a to the projectors 100 b to 100 d , which are slave devices, through the network IF 105 a.
- the CPU 101 a transmits the first position information to the projector having the IP address “192.168.254.1”. Further, in S 306 , the CPU 101 a transmits the second position information to the projector having the IP address “192.168.254.2”. Further, in S 306 , the CPU 101 a transmits the third position information to the projector having the IP address “192.168.254.3”. Further, in S 306 , the CPU 101 a transmits the fourth position information to the projector having the IP address “192.168.254.4”. After all the position information recorded in the RAM 103 a is transmitted, the CPU 101 a performs S 307 .
- the CPU 101 a controls the projection unit 111 a so as to project an image 700 a .
- the image 700 a is an image for notifying of the display pattern selected in S 303 and the projection position of the projector 100 a.
- the image 700 a in FIG. 7 is projected onto the screen by the projection unit 111 a .
- the image 700 a includes an image representing the selected display pattern 402 and information (diagonal lines in the image 700 a ) indicating that the projection position of the projector 100 a is a position corresponding to the identification information “1”.
- the projector 100 b When the projector 100 b receives the position information transmitted from the projector 100 a in S 306 , the projector 100 b performs S 322 .
- the CPU 101 b controls the image processing unit 104 b so as to generate an image for notifying a position indicated by the position information received from the projector 100 a .
- the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a.
- the projector 100 b receives the second position information from the projector 100 a . Therefore, in S 322 , an image 700 b is projected on the screen by the projection unit 111 b .
- the image 700 b is an image for notifying of a position indicated by the second position information.
- the image 700 b includes information indicating the identification information “2” (“2” in the image 700 b ) included in the second position information.
- the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b.
- the projector 100 c receives the third position information from the projector 100 a . Therefore, in S 322 , an image 700 c is projected on the screen by the projection unit 111 c .
- the image 700 c is an image for notifying of a position indicated by the third position information.
- the image 700 c includes information indicating the identification information “3” (“3” in the image 700 c ) included in the third position information.
- the projector 100 d receives the fourth position information from the projector 100 a . Therefore, in S 322 , an image 700 d is projected on the screen by the projection unit 111 d .
- the image 700 d is an image for notifying of a position indicated by the fourth position information.
- the image 700 d includes information indicating the identification information “4” (“4” in the image 700 d ) included in the fourth position information.
- the images 700 a to 700 d are displayed on the screen as illustrated in FIG. 7 .
- the user can identify the display pattern that is currently set and the projection position of the projector 100 a .
- the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
- the display pattern displayed in the image 700 a corresponds to the positions notified by the images 700 b to 700 d . Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
- FIG. 8A illustrates a case in which the IP address of the projector 100 c is inputted in the input field 522 b and the IP address of the projector 100 b is inputted in the input field 522 c .
- images 800 a to 800 d are displayed on the screen as illustrated in FIG. 8B .
- the image 800 a is an image projected onto the screen by the projection unit 111 a in S 307 when the input of the IP addresses is as illustrated in FIG. 8A .
- the image 800 a is the same image as the image 700 a .
- the image 800 b is an image projected onto the screen by the projection unit 111 b in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
- the image 800 c is an image projected onto the screen by the projection unit 111 c in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
- the image 800 d is an image projected onto the screen by the projection unit 111 d in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
- the user can identify the display pattern that is currently set and the projection position of the projector 100 a . Further, by checking the images 800 a to 800 d , the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
- the images 800 a to 800 d as illustrated in FIG. 8B are displayed on the screen, the display pattern displayed in the image 800 a does not correspond to the positions notified by the images 800 b to 800 d . Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are not correctly associated with each other.
- the user can recognize that there is an error in the input of the IP address of the projector 100 b and the IP address of the projector 100 c .
- the user can correct the error in the input of the IP addresses by using at least either one of the operation key 107 a and the remote control unit 109 a.
- a correction instruction to correct the input of the IP addresses is inputted into the CPU 101 a through the user IF 106 a .
- the CPU 101 a determines whether or not the correction instruction is inputted through the user IF 106 a .
- the CPU 101 a performs S 304 to S 307 again.
- the CPU 101 a performs S 309 .
- the CPU 101 a determines that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other and performs S 309 .
- the user can issue a projection instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a .
- the projection instruction is an instruction to start processing to display an image on the screen by the multiple projection system.
- the CPU 101 a determines whether or not the projection instruction is inputted through the user IF 106 a .
- the CPU 101 a performs S 310 .
- the CPU 101 a returns to S 308 .
- the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted into the input image 500 .
- the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
- the projector 100 b transmits the resolution information representing the resolution of the partial image projected by the projector 100 b to the projector 100 a through the network IF 105 b .
- the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 323 .
- the CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a .
- the CPU 101 a requests the projectors 100 b to 100 d to transmit area information representing a position of an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a.
- the overlapping area of the projector 100 b is a part of the image projected by the projector 100 b and the part overlaps with a part of images projected by projectors other than the projector 100 b .
- the size and the position of the overlapping area of the projector 100 b is set by the user.
- the light reduction processing is performed on the overlapping area of the projector 100 b by the image processing unit 104 b.
- the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located upward.
- the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located rightward.
- the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located downward.
- the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located leftward.
- the CPU 101 b transmits the area information indicating the position of the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a .
- the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 324 .
- the CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
- the CPU 101 a When all the resolution information and all the area information are acquired, the CPU 101 a performs S 311 .
- the CPU 101 a calculates the size of the entire image displayed on the screen by using the resolution information of the projectors 100 a to 100 d and the area information of the projectors 100 a to 100 d .
- the entire image is an image obtained by combining the partial image 210 a , the partial image 210 b , the partial image 210 c , and the partial image 210 d , which are illustrated in FIG. 1 .
- the CPU 101 a calculates coordinates indicating an area of the partial image projected by each of the projectors 100 a to 100 d in the image 200 based on the size of the entire image, the ratio of the image 200 , and the display pattern selected in S 303 .
- the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 a as first coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 b as second coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 c as third coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 d as fourth coordinate information. The CPU 101 a stores the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information, which are acquired, into the RAM 103 a . When the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information are calculated, the CPU 101 a performs S 312 .
- the CPU 101 a transmits coordinate information and the image 200 to the projectors 100 b to 100 d through the network IF 105 a.
- the CPU 101 a transmits the second coordinate information and the image 200 to the projector 100 b . Further, the CPU 101 a transmits the third coordinate information and the image 200 to the projector 100 c . Further, the CPU 101 a transmits the fourth coordinate information and the image 200 to the projector 100 d . After the second coordinate information, the third coordinate information, and the fourth coordinate information are transmitted, the CPU 101 a performs S 313 .
- the CPU 101 a reads the first coordinate information from the RAM 103 a . Thereafter, the CPU 101 a controls the image processing unit 104 a so as to generate the partial image 210 a , which is an image to be projected by the projection unit 111 a , based on the image 200 and the first coordinate information. Further, the CPU 101 a controls the image processing unit 104 a so as to perform the light reduction processing on the overlapping area in the partial image 210 a generated by the image processing unit 104 a . Then, the CPU 101 a controls the projection unit 111 a so as to project the partial image 210 a on which the light reduction processing has been performed.
- the projection unit 111 a projects the partial image 210 a , so that the partial image 210 a is displayed on the screen as illustrated in FIG. 1 .
- the CPU 101 a performs S 314 .
- the user can issue a projection end instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a .
- the projection end instruction is an instruction to end the processing to display an image on the screen by the multiple projection system.
- the CPU 101 a determines whether or not the projection end instruction is inputted.
- the CPU 101 a ends the projection processing.
- the CPU 101 a returns to S 312 and the projection unit 111 a continuously performs projection of the partial image 210 a.
- the CPU 101 b controls the image processing unit 104 b so as to generate the partial image 210 b , which is an image to be projected by the projection unit 111 b , based on the image 200 and the second coordinate information. Further, the CPU 101 b controls the image processing unit 104 b so as to perform the light reduction processing on the overlapping area in the partial image 210 b generated by the image processing unit 104 b . Then, the CPU 101 b controls the projection unit 111 b so as to project the partial image 210 b on which the light reduction processing has been performed.
- the projection unit 111 b projects the partial image 210 b , so that the partial image 210 b is displayed on the screen as illustrated in FIG. 1 .
- the CPU 101 b performs S 326 .
- the projectors 100 c and 100 d also perform, in S 325 , the same operation as that performed by the projector 100 b.
- the user can issue the projection end instruction to the projector 100 b by using at least either one of the operation key 107 b and the remote control unit 109 b . Therefore, in S 326 , the CPU 101 b determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S 326 ), the CPU 101 b ends the projection processing. When the projection end instruction is not inputted (No in S 326 ), the CPU 101 b returns to S 325 and the projection unit 111 b continuously performs projection of the partial image 210 b.
- the projectors 100 c and 100 d also perform, in S 326 , the same operation as that performed by the projector 100 b.
- S 301 to S 314 are processing realized when the CPU 101 a of the projector 100 a , which is the master device, executes a computer program stored in the ROM 102 a .
- S 321 to S 326 are processing realized when the CPU 101 b of the projector 100 b , which is the slave device, executes a computer program stored in the ROM 102 b .
- S 321 to S 326 are processing realized when the CPU 101 c of the projector 100 c , which is the slave device, executes a computer program stored in the ROM 102 c .
- S 321 to S 326 are processing realized when the CPU 101 d of the projector 100 d , which is the slave device, executes a computer program stored in the ROM 102 d.
- the display patterns that can be selected by using the selection image 400 are the display patterns 401 to 404 .
- the display patterns are not limited to these patterns.
- the display patterns that can be selected by using the selection image 400 may further include display patterns other than the display patterns 401 to 404 .
- the master device projects an image for notifying of the display pattern selected by the user and the projection position of the image to be projected by the master device.
- the slave device projects an image for notifying of the projection position of the image to be projected by the slave device.
- the user can confirm the selected display pattern and the projection positions of the images to be projected by the slave devices that are currently set. For example, when the positions indicated by the images projected by the slave devices do not correspond to the positions illustrated in the display pattern included in the image projected by the master device, the user can determine that there is an error in the input of the IP addresses of the slave devices. In this case, the user can easily determine that the positions associated with the IP addresses of the slave devices do not correspond to the current installation positions of the slave devices, so that the user can correctly modify the input of the IP addresses of the slave devices.
- the multiple projection system of the first embodiment it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
- the projector 100 a transmits position information to the projectors 100 b to 100 d and transmits information representing the display pattern selected in S 303 to the projectors 100 b to 100 d . Further, in S 307 , the projector 100 a controls the projection unit 111 a so as to project the image 900 a instead of the image 700 a . In S 322 , the projector 100 b controls the image processing unit 104 b so as to generate the image 900 b based on the position information and the information representing the display pattern which are received from the projector 100 a .
- the projector 100 b controls the projection unit 111 b so as to project the image 900 b based on the position information received from the projector 100 a .
- the projector 100 c controls the projection unit 111 c so as to project the generated image 900 c based on the position information and the information representing the display pattern which are received from the projector 100 a .
- the projector 100 d controls the projection unit 111 d so as to project the generated image 900 d based on the position information and the information representing the display pattern which are received from the projector 100 a.
- FIG. 10 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the second embodiment.
- S 301 to S 305 , S 307 to S 314 , S 321 , S 325 , and S 326 in the projection processing in FIG. 10 are the same processing as S 301 to S 305 , S 307 to S 314 , S 321 , S 325 , and S 326 in the projection processing in FIG. 3 , so that the description thereof will be omitted.
- the projection processing illustrated in FIG. 10 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices. Further, the projection processing illustrated in FIG. 10 will be described by using an example in which the display pattern 402 is selected in S 303 .
- the CPU 101 a After performing S 305 , the CPU 101 a performs S 1001 .
- the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted by the user in S 304 .
- the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
- the CPU 101 b transmits the resolution information representing the resolution of the image projected by the projector 100 b to the projector 100 a through the network IF 105 b .
- the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 1011 .
- the CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a .
- the CPU 101 a requests the projectors 100 b to 100 d to transmit area information indicating an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
- the CPU 101 b transmits the area information indicating the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a .
- the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 1012 .
- the CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
- the CPU 101 a When all the resolution information and all the area information are acquired, the CPU 101 a performs S 1002 . In S 1002 , the CPU 101 a performs determination processing to determine whether or not the setting performed in S 305 is correct by using all the acquired area information.
- the CPU 101 a performs first processing, second processing, third processing, and fourth processing.
- the first processing is processing for determining whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not.
- the CPU 101 a acquires first information indicating the position of the overlapping area included in the predetermined area 510 a from the RAM 103 a .
- the first information is stored in the RAM 103 a in advance.
- the right side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 b , so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located rightward.
- the lower side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 c , so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located downward.
- the CPU 101 a compares the first information with the area information acquired from the projector having the IP address inputted into the input field 522 a , and determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not according to the comparison result.
- the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct.
- the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is not correct.
- the second processing is processing for determining whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not.
- the CPU 101 a acquires second information indicating the position of the overlapping area included in the predetermined area 510 b from the RAM 103 a .
- the second information is stored in the RAM 103 a in advance.
- the left side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 a , so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located leftward.
- the lower side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 d , so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located downward.
- the CPU 101 a compares the second information with the area information acquired from the projector having the IP address inputted into the input field 522 b , and determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not according to the comparison result.
- the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct.
- the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct.
- the third processing is processing for determining whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not.
- the CPU 101 a acquires third information indicating the position of the overlapping area included in the predetermined area 510 c from the RAM 103 a .
- the third information is stored in the RAM 103 a in advance.
- the upper side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 a , so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located upward.
- the right side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 d , so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located rightward.
- the CPU 101 a compares the third information with the area information acquired from the projector having the IP address inputted into the input field 522 c , and determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not according to the comparison result.
- the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct.
- the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is not correct.
- the fourth processing is processing for determining whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not.
- the CPU 101 a acquires fourth information indicating the position of the overlapping area included in the predetermined area 510 d from the RAM 103 a .
- the fourth information is stored in the RAM 103 a in advance.
- the upper side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 b , so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located upward.
- the left side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 c , so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located leftward.
- the CPU 101 a compares the fourth information with the area information acquired from the projector having the IP address inputted into the input field 522 d , and determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not according to the comparison result.
- the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct.
- the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is not correct.
- the CPU 101 a determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not. After S 1002 is performed, the CPU 101 a performs S 1003 .
- the CPU 101 a transmits result information indicating the determination result of the determination processing in S 1002 and the position information recorded in the RAM 103 a to the projectors 100 b to 100 d through the network IF 105 a.
- S 1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct will be described.
- the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b .
- the result information is information indicating that there is an error in the input of the IP address.
- S 1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct will be described.
- the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b .
- the result information is information indicating that the input of the IP address is correct.
- the CPU 101 a performs S 307 .
- the projector 100 b When the projector 100 b receives the result information and the position information transmitted from the projector 100 a in S 1002 , the projector 100 b performs S 1013 .
- the CPU 101 b controls the image processing unit 104 b so as to generate an image based on the position information and the result information received from the projector 100 a.
- the image processing unit 104 b When the result information received from the projector 100 a indicates that there is an error in the input of the IP address, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information and for notifying that there is an error in the input of the IP address. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a . In this case, for example, as illustrated in FIG. 11 , the projector 100 b projects an image 1101 onto the screen. The image 1101 includes information 1102 indicating that there is an error in the input of the IP address.
- the image processing unit 104 b When the result information received from the projector 100 a indicates that the input of the IP address is correct, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a .
- the projector 100 b may control the projection unit 111 b so as to project an image indicating that the input of the IP address is correct. In this case, the projection unit 111 b projects an image indicating that the input of the IP address is correct at a position indicated by the position information received from the projector 100 a.
- the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b .
- the CPU 101 b performs S 325 .
- S 1001 and S 1002 are processing realized when the CPU 101 a of the projector 100 a , which is the master device, executes a computer program stored in the ROM 102 a .
- S 1011 to S 1013 are processing realized when the CPU 101 b of the projector 100 b , which is the slave device, executes a computer program stored in the ROM 102 b .
- S 1011 to S 1013 are processing realized when the CPU 101 c of the projector 100 c , which is the slave device, executes a computer program stored in the ROM 102 c .
- S 1011 to S 1013 are processing realized when the CPU 101 d of the projector 100 d , which is the slave device, executes a computer program stored in the ROM 102 d.
- the master device projects an image for notifying of the display pattern selected by the user and the position of the image to be projected by the master device.
- the slave device projects an image for notifying of the position of the image to be projected by the slave device.
- the master device determines whether or not there is an error in the input of the IP address of the slave device based on the position of the overlapping area of the image projected by the slave device and notifies the slave device of the determination result. Further, when the slave device is notified by the master device that there is an error in the input of the IP address, the slave device projects an image for notifying information indicating that there is an error in the input of the IP address. Thereby, the user can easily determine that the position associated with the IP address of the slave device does not correspond to the current installation position of the slave device, so that the user can correctly modify the input of the IP address of the slave device.
- the multiple projection system of the second embodiment it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
- the projector 100 a notifies the projectors 100 b to 100 d of the determination result of the determination processing of S 1002 and the projectors 100 b to 100 d project the determination result of the determination processing of S 1002 which is notified from the master device.
- the projector 100 a notifies the projectors 100 b to 100 d of the determination result of the determination processing of S 1002 and the projectors 100 b to 100 d project the determination result of the determination processing of S 1002 which is notified from the master device.
- it is not limited to this.
- the projector 100 a searches for information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b from among the first information, the second information, the third information, and the fourth information. Thereafter, the projector 100 a changes the setting of the IP address corresponding to the predetermined area 510 b according to the information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b .
- the projector 100 a searches for information corresponding to the area information acquired from the projector 100 b from among the first information, the second information, the third information, and the fourth information.
- the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct and then performs S 1003 .
- the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct. Thereafter, the projector 100 a changes the setting of the IP address of the projector 100 b so that the IP address of the projector 100 b is inputted into the input field 522 c . Thereby, the predetermined area 510 c and the IP address of the projector 100 b inputted into the input field 522 c is associated with each other.
- the projector 100 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct and performs S 1003 .
- the projector 100 a may also perform the same processing for the other predetermined areas 510 a , 510 c , and 510 d.
- a multiple projection system in the third embodiment will be described.
- the description of the same configuration and processing as those in the first and the second embodiments will be omitted.
- a multiple projection system will be described where when each of the projectors 100 a to 100 d projects a partial image based on, for example, the display pattern 402 , the display pattern is changed to a different display pattern by a user.
- the CPU 101 a performs change processing illustrated in FIG. 12 .
- the CPU 101 a When the user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the display pattern is changed, the CPU 101 a performs S 1201 .
- the CPU 101 a receives a change instruction of the display pattern through the user IF 106 a . After S 1201 is performed, the CPU 101 a performs S 1202 . In S 1202 , the CPU 101 a updates the position information set in S 305 based on the display pattern after change indicated by the change instruction of the display pattern and the display pattern before change.
- the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 401 .
- the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 403 .
- the identification information is “1” and “2” from the left (the first column) of the first row and “3” and “4” from the left (the first column) of the second row.
- the identification information is “1”, “2”, and “3” from the left (the first column) of the first row and “4”, “5”, and “6” from the left (the first column) of the second row. Therefore, the CPU 101 a updates the position information stored in the RAM 103 a so that the display pattern after change corresponds to the installation positions of the projectors 100 a to 100 d.
- the CPU 101 a associates the IP address associated with the identification information “3” (the second row, the first column) of the display pattern 402 with the identification information “4” (the first row, the third column) of the display pattern 403 and stores the IP address into the RAM 103 a . Further, the CPU 101 a associates the IP address associated with the identification information “4” (the second row, the second column) of the display pattern 402 with the identification information “5” (the second row, the first column) of the display pattern 403 and stores the IP address into the RAM 103 a .
- the CPU 101 a connects to a projector through the network IF 105 a by using the IP address corresponding to the identification information “6”. If the CPU 101 a cannot establish a connection with the projector by using the IP address corresponding to the identification information “6”, it is assumed that no projector to be a connection destination is installed.
- the CPU 101 a After S 1202 is performed, the CPU 101 a performs S 1203 .
- the CPU 101 a requests resolution information from the projectors 100 b to 100 d .
- the CPU 101 a acquires all the resolution information from the projectors 100 b to 100 d .
- the CPU 101 a performs S 1204 .
- the CPU 101 a determines the resolution of a projector of which resolution information cannot be acquired to be the same as the resolution of another projector. For example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b , the CPU 101 a determines that the resolution of the projector 100 b is the same as the resolution of the projector 100 c adjacent to the projector 100 b . Further, for example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b , the CPU 101 a may determine that the resolution of the projector 100 b is the same as the resolution of the projector 100 a which is the master device.
- the CPU 101 a requests area information from the projectors 100 b to 100 d .
- the CPU 101 a acquires all the area information from the projectors 100 b to 100 d .
- the CPU 101 a performs S 1205 .
- the CPU 101 a determines whether or not inconsistency occurs between adjacent sides based on the area information acquired in S 1204 .
- the adjacent sides are, for example, the right side of the image displayed in the area corresponding to the identification information “1” and the left side of the image displayed in the area corresponding to the identification information “2” in the display pattern 402 .
- the CPU 101 a determines that inconsistency occurs between adjacent sides (Yes in S 1205 ), the CPU 101 a performs S 1206 .
- the CPU 101 a determines that no inconsistency occurs between adjacent sides (No in S 1205 )
- the CPU 101 a performs S 1207 .
- the CPU 101 a deletes the area information acquired in S 324 from the RAM 103 a and stores the area information acquired in S 1204 into the RAM 103 a .
- the S 1206 is performed, the area information is updated.
- the CPU 101 a does not perform the processing of S 1206 and maintains consistency between the projectors in the multiple projection system by using the area information acquired in S 324 .
- FIG. 13A is a diagram illustrating an image 1301 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 401 .
- FIG. 13B is a diagram illustrating an image 1302 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 403 .
- the master device when the display pattern is changed by the user, the master device resets the projection positions of the slave devices based on the display pattern before change and the display pattern after change.
- the multiple projection system in the third embodiment has the same effect as that of the first and the second embodiments regarding the same configuration and processing as those of the first and the second embodiments.
- the master device is the projector 100 a and the slave devices are the projectors 100 b to 100 d .
- the master device may be the projector 100 d and the slave devices may be the projectors 100 a to 100 c.
- Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s).
- computer executable instructions e.g., one or more programs
- a storage medium which may also be referred to more fully as a
- the computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions.
- the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
- the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Geometry (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Controls And Circuits For Display Device (AREA)
- Transforming Electric Information Into Light Information (AREA)
Abstract
Description
- This application is a divisional, and claims the benefit, of U.S. patent application Ser. No. 15/844,352, presently pending and filed on Dec. 15, 2017, which is a continuation, and claims the benefit, of U.S. patent application Ser. No. 15/084,325, filed on Mar. 29, 2016 and issued as U.S. Pat. No. 9,875,074 on Jan. 23, 2018, and claims the benefit of, and priority to, Japanese Patent Application No. 2015-076125, filed Apr. 2, 2015, which applications and patent(s) are hereby incorporated by reference herein in their entireties.
- The present disclosure relates to a projection device and a control method therefor.
- Conventionally, a multiple projection system is known which displays one projection image by synthesizing, on a screen, a plurality of images projected onto the screen by each of a plurality of projectors.
- It is known that, in the multiple projection system, one projector (a master) of a plurality of projectors controls the other projectors (slaves).
- For example, Japanese Patent Application Laid-Open No. 2006-338044 discloses that, in the multiple projection system, one projector (a master) of a plurality of projectors associates identification information of the other projectors (slaves) with positional relationship of the other projectors (slaves). Further, it is known that the projector (master) notifies a position in one projection image displayed on a screen by the multiple projection system, to which a position of an image projected by another projector (slave) corresponds, to another projector (slave).
- However, conventionally, a method that causes a user to easily determine whether the association between the identification information of another projector (slave) and the positional relationship of another projector (slave) is correct or not is not considered. Therefore, in a state in which the association between the identification information of another projector (slave) and the positional relationship of another projector (slave) is not correct, a user may cause all the projectors of the multiple projection system to project an image. Thereby, there is a possibility that a desired projection image is not displayed on a screen at a desired timing in the multiple projection system.
- At least one aspect of the present disclosure is to solve all or at least one of the issues described above.
- In an aspect of the present disclosure, a projection device includes a projection unit configured to project an image, a communication unit configured to communicate with an external projection device, and an input unit configured to input an association between the external projection device and a predetermined area included in a display pattern for displaying one image on a screen by using an image projected by the projection device and an image projected by the external projection device. In the projection device, the projection unit projects an image for notifying of the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, and the communication unit transmits information related to the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, to the external projection device.
- According to other aspects of the present disclosure, one or more additional projection devices, one or more control methods therefor and one or more mediums are discussed herein. Further features of the present disclosure will become apparent from the following description of exemplary embodiments (with reference to the attached drawings).
- The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate exemplary embodiments, features, and aspects of the disclosure and, together with the description, serve to explain the principles of the disclosure.
-
FIG. 1 is a diagram illustrating an example of a multiple projection system in first to third embodiments. -
FIGS. 2A and 2B are diagrams illustrating an example of a configuration of a projector in the first to the third embodiments. -
FIG. 3 is a sequence diagram illustrating an example of projection processing performed in the first embodiment. -
FIG. 4 is a diagram illustrating an example of a selection image in the first to the third embodiments. -
FIGS. 5A and 5B are diagrams illustrating an example of an input image in the first to the third embodiments. -
FIG. 6 is a diagram illustrating an example of an input image in the first to the third embodiments. -
FIG. 7 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment. -
FIGS. 8A and 8B are diagrams illustrating an example of an input image in the first to the third embodiments. -
FIG. 9 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment. -
FIG. 10 is a sequence diagram illustrating an example of projection processing in the second embodiment. -
FIG. 11 is a diagram illustrating an example of an image projected by the multiple projection system in the second embodiment. -
FIG. 12 is a flowchart illustrating an example of change processing in the third embodiment. -
FIGS. 13A and 13B are diagrams illustrating an example of an image projected by the multiple projection system in the third embodiment. - Various exemplary embodiments, features, and aspects of the disclosure will be described in detail below with reference to the drawings.
-
FIG. 1 is a diagram illustrating a multiple projection system in the first embodiment. The multiple projection system is a system that displays one image on a screen by synthesizing, on the screen, images projected onto the screen by each of a plurality of projectors. - The multiple projection system of the first embodiment includes four
projectors 100 a to 100 d. Theprojectors 100 a to 100 d respectively project fourpartial images 210 a to 210 d, which are obtained by dividing oneimage 200 to be displayed on a screen into four images, onto the screen. Thereby, thepartial images 210 a to 210 d are displayed on the screen and one image corresponding to theimage 200 is displayed on the screen by combining the fourpartial images 210 a to 210 d. Thepartial image 210 a may include at least one of a part of thepartial image 210 b, a part of thepartial image 210 c, and a part of thepartial image 210 d. In the same manner, thepartial image 210 b may include at least one of a part of thepartial image 210 a, a part of thepartial image 210 c, and a part of thepartial image 210 d. In the same manner, thepartial image 210 c may include at least one of a part of thepartial image 210 a, a part of thepartial image 210 b, and a part of thepartial image 210 d. In the same manner, thepartial image 210 d may include at least one of a part of thepartial image 210 a, a part of thepartial image 210 b, and a part of thepartial image 210 c. - In the first embodiment, a case where each of the
projectors 100 a to 100 d projects a partial image obtained by dividing one image is described as an example. However, an image projected onto the screen by each of theprojectors 100 a to 100 d may be an image other than a partial image obtained by dividing one image. The projector in the first embodiment may be rephrased to a “projection device” or may be rephrased to a “display device”. - The number of projectors included in the multiple projection system is not limited to four, but may be two. The number of projectors included in the multiple projection system may be two or more.
- In the description of the first embodiment, it is assumed that the
projector 100 a of theprojectors 100 a to 100 d is set as a master device and theother projectors 100 b to 100 d are set as slave devices. In the multiple projection system, theprojector 100 a transmits theimage 200 to theprojectors 100 b to 100 d. Further, each of theprojectors 100 b to 100 d generates a partial image from theimage 200 and projects the generated partial image onto the screen. For example, theprojector 100 b generates apartial image 210 b from theimage 200 received from theprojector 100 a and projects thepartial image 210 b onto the screen. Further, for example, theprojector 100 c generates apartial image 210 c from theimage 200 received from theprojector 100 a and projects thepartial image 210 c onto the screen. Further, for example, theprojector 100 d generates apartial image 210 d from theimage 200 received from theprojector 100 a and projects thepartial image 210 d onto the screen. -
FIGS. 2A and 2B are diagrams illustrating an example of a configuration of theprojector 100 a.FIG. 2A is a diagram illustrating an example of a configuration of theprojector 100 a andFIG. 2B is an external view illustrating an example of theprojector 100 a. - The
projector 100 a includes a CPU (Central Processing Unit) 101 a, aROM 102 a, aRAM 103 a, animage processing unit 104 a, a network IF (Interface) 105 a, and a user IF 106 a. Further, theprojector 100 a includes an operation key 107 a, a remote control light-receivingunit 108 a, aprojection unit 111 a, aninput unit 112 a, and aninternal bus 113 a. - The
CPU 101 a is a control unit that controls each unit of theprojector 100 a. TheROM 102 a stores a computer program to be executed by theCPU 101 a. TheRAM 103 a temporarily stores a computer program and data as a work memory. TheCPU 101 a controls theprojector 100 a by executing the program stored in theROM 102 a. - The
image processing unit 104 a performs image processing to generate an image to be projected by theprojection unit 111 a. For example, theimage processing unit 104 a performs light reduction processing (gradation processing) on an image that overlaps a part of an image projected by a projector other than theprojector 100 a in an image projected by theprojection unit 111 a. The image that overlaps a part of an image projected by a projector other than theprojector 100 a is referred to as an “overlapping area”. - The network IF 105 a communicates with an external projector (an external projection device) through a LAN (Local Area Network). The network IF 105 a may communicate with an external projector through a wireless LAN or may communicate with an external projector through a wired LAN. The network IF 105 a may communicate with an external projector based on a communication protocol other than LAN.
- The user IF 106 a receives an input by a user from at least either one of the operation key 107 a and a
remote control unit 109 a. The remote control light-receivingunit 108 a supplies a signal received from theremote control unit 109 a to the user IF 106 a. - The
projection unit 111 a projects an image supplied from theimage processing unit 104 a. Theprojection unit 111 a includes a liquid crystal panel and a driver that drives the liquid crystal panel. Further, theprojection unit 111 a includes a drive system such as a lens and a zoom and a control system that controls a light source and a position of an image to be projected. Theinput unit 112 a receives an image from outside. Theinternal bus 113 a connects units included in theprojector 100 a. Theremote control unit 109 a is used to control theprojector 100 a. - While the
projector 100 a has been described, theprojectors 100 b to 100 d also have the same configuration as that of theprojector 100 a. For example, in the same manner as theprojector 100 a, theprojector 100 b includes a CPU 101 b, a ROM 102 b, a RAM 103 b, an image processing unit 104 b, a network IF 105 b, and a user IF 106 b. Further, theprojector 100 b includes an operation key 107 b, a remote control light-receiving unit 108 b, a projection unit 111 b, an input unit 112 b, and an internal bus 113 b. For example, in the same manner as theprojector 100 a, theprojector 100 c includes a CPU 101 c, a ROM 102 c, a RAM 103 c, an image processing unit 104 c, a network IF 105 c, and a user IF 106 c. Further, theprojector 100 c includes an operation key 107 c, a remote control light-receiving unit 108 c, a projection unit 111 c, an input unit 112 c, and an internal bus 113 c. For example, in the same manner as theprojector 100 a, theprojector 100 d includes a CPU 101 d, a ROM 102 d, a RAM 103 d, an image processing unit 104 d, a network IF 105 d, and a user IF 106 d. Further, theprojector 100 d includes an operation key 107 d, a remote control light-receiving unit 108 d, a projection unit 111 d, an input unit 112 d, and an internal bus 113 d. - Next, an operation of the
projector 100 a will be described. Theprojector 100 a supplies an image received through at least either one of theinput unit 112 a and the network IF 105 a to theimage processing unit 104 a. Thereafter, theprojector 100 a controls theprojection unit 111 a so as to project an image on which image processing is performed by theimage processing unit 104 a onto a screen. While the operation of theprojector 100 a has been described, theprojectors 100 b to 100 d also perform the same operation as that performed by theprojector 100 a. -
FIG. 3 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the first embodiment. The projection processing illustrated inFIG. 3 will be described by using an example in which theprojector 100 a is set as a master device and theprojectors 100 b to 100 d are set as slave devices. - When a user operates at least either one of the operation key 107 a and the
remote control unit 109 a and thereby theprojector 100 a is selected as the master device, theCPU 101 a performs S301. In S301, theCPU 101 a receives a master selection instruction through the user IF 106 a. Thereafter, theCPU 101 a sets an operation mode of theprojector 100 a to a master mode in which theprojector 100 a operates as the master device. In this case, theCPU 101 a stores a value indicating that theprojector 100 a is set to the master device into theRAM 103 a. After S301 is performed, theCPU 101 a performs S302. - When the user operates at least either one of the operation key 107 b and the remote control unit 109 b in the
projector 100 b and thereby theprojector 100 b is selected as the slave device, the CPU 101 b performs S321. In S321, the CPU 101 b receives a slave selection instruction through the user IF 106 b. Thereafter, the CPU 101 b sets an operation mode of theprojector 100 b to a slave mode in which theprojector 100 b operates as the slave device. In this case, the CPU 101 b stores a value indicating that theprojector 100 b is set to the slave device into the RAM 103 b. Theprojectors projector 100 b in S321. - In S302, the
CPU 101 a controls theprojection unit 111 a so as to project aselection image 400. Thereafter, theprojection unit 111 a projects theselection image 400 onto a screen. Theselection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system. -
FIG. 4 is a diagram illustrating an example of theselection image 400. Theselection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system from amongdisplay patterns 401 to 404. Thedisplay patterns 401 to 404 are images which respectively represent display patterns different from each other. - The
display pattern 401 is an image representing a pattern in which two partial images are arranged in the horizontal direction and displayed on the screen by two projectors. Further, in thedisplay pattern 401, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the right partial image. - The
display pattern 402 is an image representing a pattern of 2×2. Thedisplay pattern 402 is an image representing a pattern in which four partial images are displayed on the screen by four projectors. Further, in thedisplay pattern 402, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the upper left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the upper right partial image. Further, for example, identification information “3” is information for identifying a position of a predetermined area of the lower left partial image. Further, for example, identification information “4” is information for identifying a position of a predetermined area of the lower right partial image. - The
display pattern 403 is an image representing a display pattern of 2×3. Thedisplay pattern 403 is an image representing a pattern in which six partial images are displayed on the screen. Further, in thedisplay pattern 403, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. - The
display pattern 404 is an image representing a display pattern of 3×3. Thedisplay pattern 404 is an image representing a pattern in which nine partial images projected on the screen by nine projectors are displayed on the screen. Further, in thedisplay pattern 403, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. - The
display patterns 401 to 404 and theselection image 400 are stored in advance in theROM 102 a of theprojector 100 a which is the master device. - In S302, the
selection image 400 is displayed on the screen by theprojection unit 111 a. In this case, a user selects any one of thedisplay patterns 401 to 404 by using at least either one of the operation key 107 a and theremote control unit 109 a. When the user selects any one of thedisplay patterns 401 to 404 by using at least either one of the operation key 107 a and theremote control unit 109 a, theCPU 101 a performs S303. - In S303, the
CPU 101 a receives a selection instruction of a display pattern through the user IF 106 a. Thereafter, theCPU 101 a selects any one of thedisplay patterns 401 to 404 according to the selection instruction from the user and records information indicating the selected display pattern in theRAM 103 a. After S303 is performed, theCPU 101 a performs S304. - Hereinafter, a case in which the
display pattern 402 is selected by the user in S303 will be described as an example. In this case, in S303, theCPU 101 a selects thedisplay pattern 402 and stores information indicating thedisplay pattern 402 in theRAM 103 a. - In S304, the
CPU 101 a controls theprojection unit 111 a so as to project aninput image 500. Further, theCPU 101 a secures an area to assign IP addresses to a plurality of pieces of identification information of the display pattern selected in S302 in theRAM 103 a. - For example, when the
display pattern 402 is selected in S303, theprojection unit 111 a projects theinput image 500 illustrated inFIG. 5A onto the screen in S304. - The
input image 500 illustrated inFIG. 5A will be described. Theinput image 500 illustrated inFIG. 5A includes anarea 510, anarea 520, and aconfirmation button 530. Thearea 510 is an area where thedisplay pattern 402 selected in S302 is displayed. Thearea 520 includes anarea 521 where identification information for identifying positions ofpredetermined areas 510 a to 510 d included in thedisplay pattern 402 is displayed and aninput field 522. Thedisplay pattern 402 displayed in thearea 510 includes thepredetermined areas 510 a to 510 d, so that pieces of identification information “1” to “4” for identifying the positions of thepredetermined areas 510 a to 510 d are displayed in thearea 521. Theinput field 522 is an input field where IP addresses of the projectors are inputted with respect to the identification information “1” to “4” displayed in thearea 521. The IP address is identification information for identifying each of theprojectors 100 a to 100 d in a network for implementing the multiple projection system. - Next, the
area 521 and theinput field 522 will be described with reference toFIG. 5B . Thearea 521 includesareas 521 a to 521 d. Theinput field 522 includes input fields 522 a to 522 d. In thearea 521 a, the identification information “1” for identifying the position of thepredetermined area 510 a is displayed. Theinput field 522 a is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of thepredetermined area 510 a. The IP address inputted into theinput field 522 a is associated with thepredetermined area 510 a and the identification information “1”. - In the
area 521 b, the identification information “2” for identifying the position of thepredetermined area 510 b is displayed. Theinput field 522 b is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of thepredetermined area 510 b. The IP address inputted into theinput field 522 b is associated with thepredetermined area 510 b and the identification information “2”. - In the
area 521 c, the identification information “3” for identifying the position of thepredetermined area 510 c is displayed. Theinput field 522 c is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of thepredetermined area 510 c. The IP address inputted into theinput field 522 c is associated with thepredetermined area 510 c and the identification information “3”. - In the
area 521 d, the identification information “4” for identifying the position of thepredetermined area 510 d is displayed. Theinput field 522 d is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of thepredetermined area 510 d. The IP address inputted into theinput field 522 d is associated with thepredetermined area 510 d and the identification information “4”. - To display an image on the screen according to the
display pattern 402, it is necessary to assign IP addresses of projectors to thepredetermined areas 510 a to 510 d. Therefore, theinput image 500 is used to cause the user to input IP addresses into the input fields 522 a to 522 d. - In S304, the
input image 500 illustrated inFIG. 5A is displayed on the screen by theprojection unit 111 a. Thereafter, the user inputs IP addresses into theinput field 522 as illustrated inFIG. 5B by using at least either one of the operation key 107 a and theremote control unit 109 a. In the first embodiment, the IP address of theprojector 100 a is “192.168.254.1” and the IP address of theprojector 100 b is “192.168.254.2”. Further, in the first embodiment, the IP address of theprojector 100 c is “192.168.254.3” and the IP address of theprojector 100 d is “192.168.254.4”. - In a state in which the
input image 500 illustrated inFIG. 5A is displayed on the screen, the user can input the IP addresses of the projectors into theinput field 522 while confirming the positions where theprojectors 100 a to 100 d are installed, thearea 510, and thearea 521. In the first embodiment, it is assumed that the IP address of theprojector 100 a, which is the master device, is inputted into theinput field 522 a in advance even when the user does not input the IP address. - For example, a case in which the state where the
projectors 100 a to 100 d are installed as illustrated inFIG. 1 will be described. In this case, the user should input the IP address of theprojector 100 b into theinput field 522 b. Further, the user should input the IP address of theprojector 100 c into theinput field 522 c. Further, the user should input the IP address of theprojector 100 d into theinput field 522 d. - Further, the
CPU 101 a monitors a subnet of the IP addresses inputted into theinput field 522. When theCPU 101 a receives input of a plurality of IP addresses of which subnets are different through the user IF 106 a, theCPU 101 a controls theprojection unit 111 a so as to project information indicating that there is an error in the input of the IP addresses onto the screen. In this case, theinformation 600 such as “Subnets are different” as illustrated inFIG. 6 is projected by theprojection unit 111 a. Thereby, the user can recognize that there is an error in the input of the IP addresses. When the user inputs the IP addresses, theCPU 101 a performs S305. - In S305, the
CPU 101 a sets projection positions of theprojectors 100 a to 100 d according to the input into theinput field 522 performed by the user. - For example, in the case of
FIG. 5B , in S305, the IP address “192.168.254.1” inputted into theinput field 522 a is associated with thepredetermined area 510 a and the identification information “1”. In this case, theCPU 101 a sets the projection position, where the projector having the IP address “192.168.254.1” projects a partial image, to the position indicated by the identification information “1”. Therefore, theCPU 101 a stores first position information including the identification information “1” into theRAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.1”. The first position information may further include information indicating thearea 510 a. - For example, in the case of
FIG. 5B , in S305, the IP address “192.168.254.2” inputted into theinput field 522 b is associated with thepredetermined area 510 b and the identification information “2”. In this case, theCPU 101 a sets the projection position, where the projector having the IP address “192.168.254.2” projects a partial image, to the position indicated by the identification information “2”. Therefore, theCPU 101 a stores second position information including the identification information “2” into theRAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.2”. The second position information may further include information indicating thepredetermined area 510 b. - For example, in the case of
FIG. 5B , in S305, the IP address “192.168.254.3” inputted into theinput field 522 c is associated with thepredetermined area 510 c and the identification information “3”. In this case, theCPU 101 a sets the projection position, where the projector having the IP address “192.168.254.3” projects a partial image, to the position indicated by the identification information “3”. Therefore, theCPU 101 a stores third position information including the identification information “3” into theRAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.3”. The third position information may further include information indicating thepredetermined area 510 c. - For example, in the case of
FIG. 5B , in S305, the IP address “192.168.254.4” inputted into theinput field 522 d is associated with thepredetermined area 510 d and the identification information “4”. In this case, theCPU 101 a sets the projection position, where the projector having the IP address “192.168.254.4” projects a partial image, to the position indicated by the identification information “4”. Therefore, theCPU 101 a stores fourth position information including the identification information “4” into theRAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.4”. The fourth position information may further include information indicating thepredetermined area 510 d. - When the user presses the
confirmation button 530 by using at least either one of the operation key 107 a and theremote control unit 109 a, theCPU 101 a performs S306. When theconfirmation button 530 is pressed, the assignment of the IP addresses by using theinput image 500 is completed. - In S306, the
CPU 101 a transmits the position information recorded in theRAM 103 a to theprojectors 100 b to 100 d, which are slave devices, through the network IF 105 a. - For example, in S306, the
CPU 101 a transmits the first position information to the projector having the IP address “192.168.254.1”. Further, in S306, theCPU 101 a transmits the second position information to the projector having the IP address “192.168.254.2”. Further, in S306, theCPU 101 a transmits the third position information to the projector having the IP address “192.168.254.3”. Further, in S306, theCPU 101 a transmits the fourth position information to the projector having the IP address “192.168.254.4”. After all the position information recorded in theRAM 103 a is transmitted, theCPU 101 a performs S307. - In S307, the
CPU 101 a controls theprojection unit 111 a so as to project animage 700 a. Theimage 700 a is an image for notifying of the display pattern selected in S303 and the projection position of theprojector 100 a. - When the
display pattern 402 is selected in S303, theimage 700 a inFIG. 7 is projected onto the screen by theprojection unit 111 a. Theimage 700 a includes an image representing the selecteddisplay pattern 402 and information (diagonal lines in theimage 700 a) indicating that the projection position of theprojector 100 a is a position corresponding to the identification information “1”. - When the
projector 100 b receives the position information transmitted from theprojector 100 a in S306, theprojector 100 b performs S322. In S322, the CPU 101 b controls the image processing unit 104 b so as to generate an image for notifying a position indicated by the position information received from theprojector 100 a. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from theprojector 100 a. - For example, when the IP address of the
projector 100 b is inputted into theinput field 522 b by the user, theprojector 100 b receives the second position information from theprojector 100 a. Therefore, in S322, animage 700 b is projected on the screen by the projection unit 111 b. Theimage 700 b is an image for notifying of a position indicated by the second position information. Theimage 700 b includes information indicating the identification information “2” (“2” in theimage 700 b) included in the second position information. - While the operation in which the
projector 100 b performs S322 has been described, theprojectors projector 100 b. - For example, when the IP address of the
projector 100 c is inputted into theinput field 522 c by the user, theprojector 100 c receives the third position information from theprojector 100 a. Therefore, in S322, animage 700 c is projected on the screen by the projection unit 111 c. Theimage 700 c is an image for notifying of a position indicated by the third position information. Theimage 700 c includes information indicating the identification information “3” (“3” in theimage 700 c) included in the third position information. - For example, when the IP address of the
projector 100 d is inputted into theinput field 522 d by the user, theprojector 100 d receives the fourth position information from theprojector 100 a. Therefore, in S322, animage 700 d is projected on the screen by the projection unit 111 d. Theimage 700 d is an image for notifying of a position indicated by the fourth position information. Theimage 700 d includes information indicating the identification information “4” (“4” in theimage 700 d) included in the fourth position information. - When S307 is performed by the
projector 100 a and S322 is performed by theprojectors 100 b to 100 d, theimages 700 a to 700 d are displayed on the screen as illustrated inFIG. 7 . In this case, by checking theimage 700 a, the user can identify the display pattern that is currently set and the projection position of theprojector 100 a. Further, by checking theimages 700 a to 700 d, the user can compare the display pattern that is currently set with the projection positions of theprojectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of theprojectors 100 b to 100 d and the positional relationship between theprojectors 100 b to 100 d are correctly associated with each other. When theimages 700 b to 700 d as illustrated inFIG. 7 are displayed on the screen, the display pattern displayed in theimage 700 a corresponds to the positions notified by theimages 700 b to 700 d. Therefore, the user can determine that the IP addresses of theprojectors 100 b to 100 d and the positional relationship between theprojectors 100 b to 100 d are correctly associated with each other. - When all the IP addresses of the
projectors 100 b to 100 d are correctly inputted in theinput field 522 as illustrated inFIG. 5B , theimages 700 b to 700 d as illustrated inFIG. 7 are displayed on the screen by performing S307 and S322. - However, there is a case in which the IP addresses of the
projectors 100 b to 100 d are not correctly inputted in theinput field 522. This case will be described with reference toFIG. 8A .FIG. 8A illustrates a case in which the IP address of theprojector 100 c is inputted in theinput field 522 b and the IP address of theprojector 100 b is inputted in theinput field 522 c. In this case, it is difficult for the user to recognize that there is an error in the input of the IP addresses of theprojector 100 b and theprojector 100 c. However, in this case, when S307 is performed by theprojector 100 a and S322 is performed by theprojectors 100 b to 100 d,images 800 a to 800 d are displayed on the screen as illustrated inFIG. 8B . Theimage 800 a is an image projected onto the screen by theprojection unit 111 a in S307 when the input of the IP addresses is as illustrated inFIG. 8A . Theimage 800 a is the same image as theimage 700 a. Theimage 800 b is an image projected onto the screen by the projection unit 111 b in S322 when the input of the IP addresses is as illustrated inFIG. 8A . Theimage 800 c is an image projected onto the screen by the projection unit 111 c in S322 when the input of the IP addresses is as illustrated inFIG. 8A . Theimage 800 d is an image projected onto the screen by the projection unit 111 d in S322 when the input of the IP addresses is as illustrated inFIG. 8A . - In this case, by checking the
image 800 a, the user can identify the display pattern that is currently set and the projection position of theprojector 100 a. Further, by checking theimages 800 a to 800 d, the user can compare the display pattern that is currently set with the projection positions of theprojectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of theprojectors 100 b to 100 d and the positional relationship between theprojectors 100 b to 100 d are correctly associated with each other. When theimages 800 a to 800 d as illustrated inFIG. 8B are displayed on the screen, the display pattern displayed in theimage 800 a does not correspond to the positions notified by theimages 800 b to 800 d. Therefore, the user can determine that the IP addresses of theprojectors 100 b to 100 d and the positional relationship between theprojectors 100 b to 100 d are not correctly associated with each other. - When the
images 800 a to 800 d as illustrated inFIG. 8B are displayed, the user can recognize that there is an error in the input of the IP address of theprojector 100 b and the IP address of theprojector 100 c. In this case, the user can correct the error in the input of the IP addresses by using at least either one of the operation key 107 a and theremote control unit 109 a. - When the user performs an operation to correct the input of the IP addresses by using at least either one of the operation key 107 a and the
remote control unit 109 a, a correction instruction to correct the input of the IP addresses is inputted into theCPU 101 a through the user IF 106 a. In S308, theCPU 101 a determines whether or not the correction instruction is inputted through the user IF 106 a. When the correction instruction is inputted (Yes in S308), theCPU 101 a performs S304 to S307 again. When the correction instruction is not inputted (No in S308), theCPU 101 a performs S309. - When the correction instruction is not inputted (No in S308), the
CPU 101 a determines that the IP addresses of theprojectors 100 b to 100 d and the positional relationship between theprojectors 100 b to 100 d are correctly associated with each other and performs S309. When the correction instruction is not inputted (No in S308), the user can issue a projection instruction to theprojector 100 a by using at least either one of the operation key 107 a and theremote control unit 109 a. The projection instruction is an instruction to start processing to display an image on the screen by the multiple projection system. - In S309, the
CPU 101 a determines whether or not the projection instruction is inputted through the user IF 106 a. When the projection instruction is inputted (Yes in S309), theCPU 101 a performs S310. When the projection instruction is not inputted (No in S309), theCPU 101 a returns to S308. - In S310, the
CPU 101 a connects to theprojectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted into theinput image 500. - After S310 is performed, the
CPU 101 a requests theprojectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of theprojectors 100 b to 100 d to theprojector 100 a. Thereafter, in S323, theprojector 100 b transmits the resolution information representing the resolution of the partial image projected by theprojector 100 b to theprojector 100 a through the network IF 105 b. Theprojectors projector 100 b in S323. TheCPU 101 a stores the resolution information received through the network IF 105 a into theRAM 103 a. Thereafter, theCPU 101 a requests theprojectors 100 b to 100 d to transmit area information representing a position of an overlapping area of an image projected by each of theprojectors 100 b to 100 d to theprojector 100 a. - The overlapping area of the
projector 100 b is a part of the image projected by theprojector 100 b and the part overlaps with a part of images projected by projectors other than theprojector 100 b. The size and the position of the overlapping area of theprojector 100 b is set by the user. The light reduction processing is performed on the overlapping area of theprojector 100 b by the image processing unit 104 b. - When the upper side of the image projected by the
projector 100 b overlaps with a part of an image projected by projectors other than theprojector 100 b, the area information of theprojector 100 b includes information indicating that the overlapping area of theprojector 100 b is located upward. When the right side of the image projected by theprojector 100 b overlaps with a part of an image projected by projectors other than theprojector 100 b, the area information of theprojector 100 b includes information indicating that the overlapping area of theprojector 100 b is located rightward. When the lower side of the image projected by theprojector 100 b overlaps with a part of an image projected by projectors other than theprojector 100 b, the area information of theprojector 100 b includes information indicating that the overlapping area of theprojector 100 b is located downward. When the left side of the image projected by theprojector 100 b overlaps with a part of an image projected by projectors other than theprojector 100 b, the area information of theprojector 100 b includes information indicating that the overlapping area of theprojector 100 b is located leftward. - In S324, the CPU 101 b transmits the area information indicating the position of the overlapping area of the
projector 100 b to theprojector 100 a through the network IF 105 b according to the request from theprojector 100 a. Theprojectors projector 100 b in S324. TheCPU 101 a stores the area information received through the network IF 105 a into theRAM 103 a. - It is assumed that the resolution information of the
projector 100 a and the area information of theprojector 100 a are stored in theRAM 103 a. - When all the resolution information and all the area information are acquired, the
CPU 101 a performs S311. In S311, theCPU 101 a calculates the size of the entire image displayed on the screen by using the resolution information of theprojectors 100 a to 100 d and the area information of theprojectors 100 a to 100 d. The entire image is an image obtained by combining thepartial image 210 a, thepartial image 210 b, thepartial image 210 c, and thepartial image 210 d, which are illustrated inFIG. 1 . Further, theCPU 101 a calculates coordinates indicating an area of the partial image projected by each of theprojectors 100 a to 100 d in theimage 200 based on the size of the entire image, the ratio of theimage 200, and the display pattern selected in S303. - For example, in S311, the
CPU 101 a calculates coordinates indicating the area of the partial image of theprojector 100 a as first coordinate information. Further, theCPU 101 a calculates coordinates indicating the area of the partial image of theprojector 100 b as second coordinate information. Further, theCPU 101 a calculates coordinates indicating the area of the partial image of theprojector 100 c as third coordinate information. Further, theCPU 101 a calculates coordinates indicating the area of the partial image of theprojector 100 d as fourth coordinate information. TheCPU 101 a stores the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information, which are acquired, into theRAM 103 a. When the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information are calculated, theCPU 101 a performs S312. - In S312, the
CPU 101 a transmits coordinate information and theimage 200 to theprojectors 100 b to 100 d through the network IF 105 a. - For example, in S312, the
CPU 101 a transmits the second coordinate information and theimage 200 to theprojector 100 b. Further, theCPU 101 a transmits the third coordinate information and theimage 200 to theprojector 100 c. Further, theCPU 101 a transmits the fourth coordinate information and theimage 200 to theprojector 100 d. After the second coordinate information, the third coordinate information, and the fourth coordinate information are transmitted, theCPU 101 a performs S313. - In S313, the
CPU 101 a reads the first coordinate information from theRAM 103 a. Thereafter, theCPU 101 a controls theimage processing unit 104 a so as to generate thepartial image 210 a, which is an image to be projected by theprojection unit 111 a, based on theimage 200 and the first coordinate information. Further, theCPU 101 a controls theimage processing unit 104 a so as to perform the light reduction processing on the overlapping area in thepartial image 210 a generated by theimage processing unit 104 a. Then, theCPU 101 a controls theprojection unit 111 a so as to project thepartial image 210 a on which the light reduction processing has been performed. In S313, theprojection unit 111 a projects thepartial image 210 a, so that thepartial image 210 a is displayed on the screen as illustrated inFIG. 1 . After the partial image is projected by theprojection unit 111 a, theCPU 101 a performs S314. The user can issue a projection end instruction to theprojector 100 a by using at least either one of the operation key 107 a and theremote control unit 109 a. The projection end instruction is an instruction to end the processing to display an image on the screen by the multiple projection system. - Therefore, in S314, the
CPU 101 a determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S314), theCPU 101 a ends the projection processing. When the projection end instruction is not inputted (No in S314), theCPU 101 a returns to S312 and theprojection unit 111 a continuously performs projection of thepartial image 210 a. - After S312 is performed, when the CPU 101 b receives the second coordinate information and the
image 200 from theprojector 100 a, the CPU 101 b performs S325. In S325, the CPU 101 b controls the image processing unit 104 b so as to generate thepartial image 210 b, which is an image to be projected by the projection unit 111 b, based on theimage 200 and the second coordinate information. Further, the CPU 101 b controls the image processing unit 104 b so as to perform the light reduction processing on the overlapping area in thepartial image 210 b generated by the image processing unit 104 b. Then, the CPU 101 b controls the projection unit 111 b so as to project thepartial image 210 b on which the light reduction processing has been performed. In S325, the projection unit 111 b projects thepartial image 210 b, so that thepartial image 210 b is displayed on the screen as illustrated inFIG. 1 . After the partial image is projected by the projection unit 111 b, the CPU 101 b performs S326. - While the operation in which the
projector 100 b performs S325 has been described, theprojectors projector 100 b. - The user can issue the projection end instruction to the
projector 100 b by using at least either one of the operation key 107 b and the remote control unit 109 b. Therefore, in S326, the CPU 101 b determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S326), the CPU 101 b ends the projection processing. When the projection end instruction is not inputted (No in S326), the CPU 101 b returns to S325 and the projection unit 111 b continuously performs projection of thepartial image 210 b. - While the operation in which the
projector 100 b performs S326 has been described, theprojectors projector 100 b. - S301 to S314 are processing realized when the
CPU 101 a of theprojector 100 a, which is the master device, executes a computer program stored in theROM 102 a. S321 to S326 are processing realized when the CPU 101 b of theprojector 100 b, which is the slave device, executes a computer program stored in the ROM 102 b. S321 to S326 are processing realized when the CPU 101 c of theprojector 100 c, which is the slave device, executes a computer program stored in the ROM 102 c. S321 to S326 are processing realized when the CPU 101 d of theprojector 100 d, which is the slave device, executes a computer program stored in the ROM 102 d. - In the first embodiment, it is described that the display patterns that can be selected by using the
selection image 400 are thedisplay patterns 401 to 404. However, the display patterns are not limited to these patterns. For example, the display patterns that can be selected by using theselection image 400 may further include display patterns other than thedisplay patterns 401 to 404. - In the first embodiment, in the multiple projection system, the master device projects an image for notifying of the display pattern selected by the user and the projection position of the image to be projected by the master device. The slave device projects an image for notifying of the projection position of the image to be projected by the slave device. Thereby, the user can confirm the selected display pattern and the projection positions of the images to be projected by the slave devices that are currently set. For example, when the positions indicated by the images projected by the slave devices do not correspond to the positions illustrated in the display pattern included in the image projected by the master device, the user can determine that there is an error in the input of the IP addresses of the slave devices. In this case, the user can easily determine that the positions associated with the IP addresses of the slave devices do not correspond to the current installation positions of the slave devices, so that the user can correctly modify the input of the IP addresses of the slave devices.
- Therefore, in the multiple projection system of the first embodiment, it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
- In the first embodiment, it is described that when S307 is performed by the master device and S322 is performed by the slave devices, in the multiple projection system, images such as the
images 700 a to 700 d as illustrated inFIG. 7 are displayed on the screen. However, it is not limited to this. For example, in the first embodiment, when S307 is performed by the master device and S322 is performed by the slave devices, in the multiple projection system, images such as theimages 900 a to 900 d as illustrated inFIG. 9 may be displayed on the screen. - In this case, in S306, the
projector 100 a transmits position information to theprojectors 100 b to 100 d and transmits information representing the display pattern selected in S303 to theprojectors 100 b to 100 d. Further, in S307, theprojector 100 a controls theprojection unit 111 a so as to project theimage 900 a instead of theimage 700 a. In S322, theprojector 100 b controls the image processing unit 104 b so as to generate theimage 900 b based on the position information and the information representing the display pattern which are received from theprojector 100 a. Further, theprojector 100 b controls the projection unit 111 b so as to project theimage 900 b based on the position information received from theprojector 100 a. In the same manner as theprojector 100 b, in S322, theprojector 100 c controls the projection unit 111 c so as to project the generatedimage 900 c based on the position information and the information representing the display pattern which are received from theprojector 100 a. In the same manner as theprojector 100 b, in S322, theprojector 100 d controls the projection unit 111 d so as to project the generatedimage 900 d based on the position information and the information representing the display pattern which are received from theprojector 100 a. - Next, a multiple projection system in the second embodiment will be described. In the second embodiment, the description of the same configuration and processing as those in the first embodiment will be omitted.
-
FIG. 10 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the second embodiment. - S301 to S305, S307 to S314, S321, S325, and S326 in the projection processing in
FIG. 10 are the same processing as S301 to S305, S307 to S314, S321, S325, and S326 in the projection processing inFIG. 3 , so that the description thereof will be omitted. - The projection processing illustrated in
FIG. 10 will be described by using an example in which theprojector 100 a is set as a master device and theprojectors 100 b to 100 d are set as slave devices. Further, the projection processing illustrated inFIG. 10 will be described by using an example in which thedisplay pattern 402 is selected in S303. - After performing S305, the
CPU 101 a performs S1001. In S1001, in the same manner as in S310 inFIG. 3 , theCPU 101 a connects to theprojectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted by the user in S304. - After S1001 is performed, the
CPU 101 a requests theprojectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of theprojectors 100 b to 100 d to theprojector 100 a. Thereafter, in the same manner as in S323 inFIG. 3 , in S1011, the CPU 101 b transmits the resolution information representing the resolution of the image projected by theprojector 100 b to theprojector 100 a through the network IF 105 b. Theprojectors projector 100 b in S1011. TheCPU 101 a stores the resolution information received through the network IF 105 a into theRAM 103 a. Thereafter, theCPU 101 a requests theprojectors 100 b to 100 d to transmit area information indicating an overlapping area of an image projected by each of theprojectors 100 b to 100 d to theprojector 100 a. In the same manner as in S324 inFIG. 3 , in S1012, the CPU 101 b transmits the area information indicating the overlapping area of theprojector 100 b to theprojector 100 a through the network IF 105 b according to the request from theprojector 100 a. Theprojectors projector 100 b in S1012. TheCPU 101 a stores the area information received through the network IF 105 a into theRAM 103 a. - When all the resolution information and all the area information are acquired, the
CPU 101 a performs S1002. In S1002, theCPU 101 a performs determination processing to determine whether or not the setting performed in S305 is correct by using all the acquired area information. - As the determination processing, the
CPU 101 a performs first processing, second processing, third processing, and fourth processing. - The first processing is processing for determining whether the association between the
predetermined area 510 a and the IP address inputted into theinput field 522 a is correct or not. TheCPU 101 a acquires first information indicating the position of the overlapping area included in thepredetermined area 510 a from theRAM 103 a. The first information is stored in theRAM 103 a in advance. The right side of the image projected onto thepredetermined area 510 a overlaps with a part of the image projected onto thepredetermined area 510 b, so that the first information includes information indicating that the overlapping area included in thepredetermined area 510 a is located rightward. Further, the lower side of the image projected onto thepredetermined area 510 a overlaps with a part of the image projected onto thepredetermined area 510 c, so that the first information includes information indicating that the overlapping area included in thepredetermined area 510 a is located downward. - The
CPU 101 a compares the first information with the area information acquired from the projector having the IP address inputted into theinput field 522 a, and determines whether the association between thepredetermined area 510 a and the IP address inputted into theinput field 522 a is correct or not according to the comparison result. When the first information corresponds with the area information acquired from the projector having the IP address inputted into theinput field 522 a, theCPU 101 a determines that the association between thepredetermined area 510 a and the IP address inputted into theinput field 522 a is correct. When the first information does not correspond with the area information acquired from the projector having the IP address inputted into theinput field 522 a, theCPU 101 a determines that the association between thepredetermined area 510 a and the IP address inputted into theinput field 522 a is not correct. - The second processing is processing for determining whether the association between the
predetermined area 510 b and the IP address inputted into theinput field 522 b is correct or not. TheCPU 101 a acquires second information indicating the position of the overlapping area included in thepredetermined area 510 b from theRAM 103 a. The second information is stored in theRAM 103 a in advance. The left side of the image projected onto thepredetermined area 510 b overlaps with a part of the image projected onto thepredetermined area 510 a, so that the second information includes information indicating that the overlapping area included in thepredetermined area 510 b is located leftward. Further, the lower side of the image projected onto thepredetermined area 510 b overlaps with a part of the image projected onto thepredetermined area 510 d, so that the second information includes information indicating that the overlapping area included in thepredetermined area 510 b is located downward. - The
CPU 101 a compares the second information with the area information acquired from the projector having the IP address inputted into theinput field 522 b, and determines whether the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is correct or not according to the comparison result. When the second information corresponds with the area information acquired from the projector having the IP address inputted into theinput field 522 b, theCPU 101 a determines that the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is correct. When the second information does not correspond with the area information acquired from the projector having the IP address inputted into theinput field 522 b, theCPU 101 a determines that the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is not correct. - The third processing is processing for determining whether the association between the
predetermined area 510 c and the IP address inputted into theinput field 522 c is correct or not. TheCPU 101 a acquires third information indicating the position of the overlapping area included in thepredetermined area 510 c from theRAM 103 a. The third information is stored in theRAM 103 a in advance. The upper side of the image projected onto thepredetermined area 510 c overlaps with a part of the image projected onto thepredetermined area 510 a, so that the third information includes information indicating that the overlapping area included in thepredetermined area 510 c is located upward. Further, the right side of the image projected onto thepredetermined area 510 c overlaps with a part of the image projected onto thepredetermined area 510 d, so that the third information includes information indicating that the overlapping area included in thepredetermined area 510 c is located rightward. - The
CPU 101 a compares the third information with the area information acquired from the projector having the IP address inputted into theinput field 522 c, and determines whether the association between thepredetermined area 510 c and the IP address inputted into theinput field 522 c is correct or not according to the comparison result. When the third information corresponds with the area information acquired from the projector having the IP address inputted into theinput field 522 c, theCPU 101 a determines that the association between thepredetermined area 510 c and the IP address inputted into theinput field 522 c is correct. When the third information does not correspond with the area information acquired from the projector having the IP address inputted into theinput field 522 c, theCPU 101 a determines that the association between thepredetermined area 510 c and the IP address inputted into theinput field 522 c is not correct. - The fourth processing is processing for determining whether the association between the
predetermined area 510 d and the IP address inputted into theinput field 522 d is correct or not. TheCPU 101 a acquires fourth information indicating the position of the overlapping area included in thepredetermined area 510 d from theRAM 103 a. The fourth information is stored in theRAM 103 a in advance. The upper side of the image projected onto thepredetermined area 510 d overlaps with a part of the image projected onto thepredetermined area 510 b, so that the fourth information includes information indicating that the overlapping area included in thepredetermined area 510 d is located upward. Further, the left side of the image projected onto thepredetermined area 510 d overlaps with a part of the image projected onto thepredetermined area 510 c, so that the fourth information includes information indicating that the overlapping area included in thepredetermined area 510 d is located leftward. - The
CPU 101 a compares the fourth information with the area information acquired from the projector having the IP address inputted into theinput field 522 d, and determines whether the association between thepredetermined area 510 d and the IP address inputted into theinput field 522 d is correct or not according to the comparison result. When the fourth information corresponds with the area information acquired from the projector having the IP address inputted into theinput field 522 d, theCPU 101 a determines that the association between thepredetermined area 510 d and the IP address inputted into theinput field 522 d is correct. When the fourth information does not correspond with the area information acquired from the projector having the IP address inputted into theinput field 522 d, theCPU 101 a determines that the association between thepredetermined area 510 d and the IP address inputted into theinput field 522 d is not correct. - When the determination processing is performed in S1002, the
CPU 101 a determines whether the association between thepredetermined area 510 a and the IP address inputted into theinput field 522 a is correct or not. Further, theCPU 101 a determines whether the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is correct or not. Further, theCPU 101 a determines whether the association between thepredetermined area 510 c and the IP address inputted into theinput field 522 c is correct or not. Further, theCPU 101 a determines whether the association between thepredetermined area 510 d and the IP address inputted into theinput field 522 d is correct or not. After S1002 is performed, theCPU 101 a performs S1003. - In S1003, the
CPU 101 a transmits result information indicating the determination result of the determination processing in S1002 and the position information recorded in theRAM 103 a to theprojectors 100 b to 100 d through the network IF 105 a. - S1003 performed by the
CPU 101 a when it is determined that, for example, the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is not correct will be described. In this case, in S1003, theCPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into theinput field 522 b. The result information is information indicating that there is an error in the input of the IP address. - Further, S1003 performed by the
CPU 101 a when it is determined that, for example, the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is correct will be described. In this case, in S1003, theCPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into theinput field 522 b. The result information is information indicating that the input of the IP address is correct. After S1003 is performed, theCPU 101 a performs S307. - When the
projector 100 b receives the result information and the position information transmitted from theprojector 100 a in S1002, theprojector 100 b performs S1013. In S1013, the CPU 101 b controls the image processing unit 104 b so as to generate an image based on the position information and the result information received from theprojector 100 a. - When the result information received from the
projector 100 a indicates that there is an error in the input of the IP address, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information and for notifying that there is an error in the input of the IP address. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from theprojector 100 a. In this case, for example, as illustrated inFIG. 11 , theprojector 100 b projects animage 1101 onto the screen. Theimage 1101 includesinformation 1102 indicating that there is an error in the input of the IP address. - When the result information received from the
projector 100 a indicates that the input of the IP address is correct, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from theprojector 100 a. When the result information received from theprojector 100 a indicates that the input of the IP address is correct, theprojector 100 b may control the projection unit 111 b so as to project an image indicating that the input of the IP address is correct. In this case, the projection unit 111 b projects an image indicating that the input of the IP address is correct at a position indicated by the position information received from theprojector 100 a. - While the operation of the
projector 100 b has been described, theprojectors projector 100 b. After S1013 is performed, the CPU 101 b performs S325. - S1001 and S1002 are processing realized when the
CPU 101 a of theprojector 100 a, which is the master device, executes a computer program stored in theROM 102 a. S1011 to S1013 are processing realized when the CPU 101 b of theprojector 100 b, which is the slave device, executes a computer program stored in the ROM 102 b. S1011 to S1013 are processing realized when the CPU 101 c of theprojector 100 c, which is the slave device, executes a computer program stored in the ROM 102 c. S1011 to S1013 are processing realized when the CPU 101 d of theprojector 100 d, which is the slave device, executes a computer program stored in the ROM 102 d. - In the multiple projection system in the second embodiment, the master device projects an image for notifying of the display pattern selected by the user and the position of the image to be projected by the master device. Further, the slave device projects an image for notifying of the position of the image to be projected by the slave device.
- In the second embodiment, the master device determines whether or not there is an error in the input of the IP address of the slave device based on the position of the overlapping area of the image projected by the slave device and notifies the slave device of the determination result. Further, when the slave device is notified by the master device that there is an error in the input of the IP address, the slave device projects an image for notifying information indicating that there is an error in the input of the IP address. Thereby, the user can easily determine that the position associated with the IP address of the slave device does not correspond to the current installation position of the slave device, so that the user can correctly modify the input of the IP address of the slave device.
- Therefore, in the multiple projection system of the second embodiment, it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
- It is assumed that the multiple projection system in the second embodiment has the same effect as that of the first embodiment regarding the same configuration and processing as those of the first embodiment.
- In the second embodiment, the
projector 100 a notifies theprojectors 100 b to 100 d of the determination result of the determination processing of S1002 and theprojectors 100 b to 100 d project the determination result of the determination processing of S1002 which is notified from the master device. However, it is not limited to this. - For example, when the
projector 100 a receives all the area information, theprojector 100 a searches for information corresponding to the area information acquired from the projector having the IP address inputted into theinput field 522 b from among the first information, the second information, the third information, and the fourth information. Thereafter, theprojector 100 a changes the setting of the IP address corresponding to thepredetermined area 510 b according to the information corresponding to the area information acquired from the projector having the IP address inputted into theinput field 522 b. For example, when the IP address of theprojector 100 b is inputted into theinput field 522 b in S305, theprojector 100 a searches for information corresponding to the area information acquired from theprojector 100 b from among the first information, the second information, the third information, and the fourth information. When the information corresponding to the area information acquired from theprojector 100 b is the second information, theprojector 100 a determines that the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is correct and then performs S1003. When the information corresponding to the area information acquired from theprojector 100 b is the third information, theprojector 100 a determines that the association between thepredetermined area 510 b and the IP address inputted into theinput field 522 b is not correct. Thereafter, theprojector 100 a changes the setting of the IP address of theprojector 100 b so that the IP address of theprojector 100 b is inputted into theinput field 522 c. Thereby, thepredetermined area 510 c and the IP address of theprojector 100 b inputted into theinput field 522 c is associated with each other. Thereafter, theprojector 100 a determines that the association between thepredetermined area 510 c and the IP address inputted into theinput field 522 c is correct and performs S1003. Theprojector 100 a may also perform the same processing for the otherpredetermined areas - Next, a multiple projection system in the third embodiment will be described. In the third embodiment, the description of the same configuration and processing as those in the first and the second embodiments will be omitted. In the third embodiment, a multiple projection system will be described where when each of the
projectors 100 a to 100 d projects a partial image based on, for example, thedisplay pattern 402, the display pattern is changed to a different display pattern by a user. After the processing of S313 described inFIG. 3 is performed, theCPU 101 a performs change processing illustrated inFIG. 12 . - When the user operates at least either one of the operation key 107 a and the
remote control unit 109 a and thereby the display pattern is changed, theCPU 101 a performs S1201. - In S1201, the
CPU 101 a receives a change instruction of the display pattern through the user IF 106 a. After S1201 is performed, theCPU 101 a performs S1202. In S1202, theCPU 101 a updates the position information set in S305 based on the display pattern after change indicated by the change instruction of the display pattern and the display pattern before change. Hereinafter, regarding the processing of S1201, a case will be described in which the display pattern before change is thedisplay pattern 402 and the display pattern after change is thedisplay pattern 401. Further, a case will be described in which the display pattern before change is thedisplay pattern 402 and the display pattern after change is thedisplay pattern 403. - In the
display pattern 402, the identification information is “1” and “2” from the left (the first column) of the first row and “3” and “4” from the left (the first column) of the second row. On the other hand, in thedisplay pattern 403, the identification information is “1”, “2”, and “3” from the left (the first column) of the first row and “4”, “5”, and “6” from the left (the first column) of the second row. Therefore, theCPU 101 a updates the position information stored in theRAM 103 a so that the display pattern after change corresponds to the installation positions of theprojectors 100 a to 100 d. - For example, the
CPU 101 a associates the IP address associated with the identification information “3” (the second row, the first column) of thedisplay pattern 402 with the identification information “4” (the first row, the third column) of thedisplay pattern 403 and stores the IP address into theRAM 103 a. Further, theCPU 101 a associates the IP address associated with the identification information “4” (the second row, the second column) of thedisplay pattern 402 with the identification information “5” (the second row, the first column) of thedisplay pattern 403 and stores the IP address into theRAM 103 a. Although no IP address is associated with the identification information “3”, if the IP address for the identification information “6” is set, theCPU 101 a connects to a projector through the network IF 105 a by using the IP address corresponding to the identification information “6”. If theCPU 101 a cannot establish a connection with the projector by using the IP address corresponding to the identification information “6”, it is assumed that no projector to be a connection destination is installed. - After S1202 is performed, the
CPU 101 a performs S1203. In S1203, theCPU 101 a requests resolution information from theprojectors 100 b to 100 d. When theCPU 101 a acquires all the resolution information from theprojectors 100 b to 100 d, theCPU 101 a performs S1204. TheCPU 101 a determines the resolution of a projector of which resolution information cannot be acquired to be the same as the resolution of another projector. For example, if theCPU 101 a cannot acquire the resolution information from theprojector 100 b, theCPU 101 a determines that the resolution of theprojector 100 b is the same as the resolution of theprojector 100 c adjacent to theprojector 100 b. Further, for example, if theCPU 101 a cannot acquire the resolution information from theprojector 100 b, theCPU 101 a may determine that the resolution of theprojector 100 b is the same as the resolution of theprojector 100 a which is the master device. - In S1204, the
CPU 101 a requests area information from theprojectors 100 b to 100 d. When theCPU 101 a acquires all the area information from theprojectors 100 b to 100 d, theCPU 101 a performs S1205. - In S1205, the
CPU 101 a determines whether or not inconsistency occurs between adjacent sides based on the area information acquired in S1204. The adjacent sides are, for example, the right side of the image displayed in the area corresponding to the identification information “1” and the left side of the image displayed in the area corresponding to the identification information “2” in thedisplay pattern 402. - When the
CPU 101 a determines that inconsistency occurs between adjacent sides (Yes in S1205), theCPU 101 a performs S1206. When theCPU 101 a determines that no inconsistency occurs between adjacent sides (No in S1205), theCPU 101 a performs S1207. In S1206, theCPU 101 a deletes the area information acquired in S324 from theRAM 103 a and stores the area information acquired in S1204 into theRAM 103 a. When the S1206 is performed, the area information is updated. - When the display pattern is changed from the
display pattern 402 to thedisplay pattern 403 by the user, the number of projectors that project partial images increases. In this case, theCPU 101 a does not perform the processing of S1206 and maintains consistency between the projectors in the multiple projection system by using the area information acquired in S324. - After S1206 is performed, the
CPU 101 a performs S1207. The processes of S1207 to S1209 are respectively the same as the processes of S311 to S313 illustrated inFIG. 3 .FIG. 13A is a diagram illustrating animage 1301 displayed on the screen when the display pattern is changed from thedisplay pattern 402 to thedisplay pattern 401.FIG. 13B is a diagram illustrating animage 1302 displayed on the screen when the display pattern is changed from thedisplay pattern 402 to thedisplay pattern 403. - In the multiple projection system in the third embodiment, when the display pattern is changed by the user, the master device resets the projection positions of the slave devices based on the display pattern before change and the display pattern after change.
- Thereby, in the multiple projection system in the third embodiment, it is possible to automatically set the projection positions of the slave devices by changing the display pattern by the user.
- It is assumed that the multiple projection system in the third embodiment has the same effect as that of the first and the second embodiments regarding the same configuration and processing as those of the first and the second embodiments.
- In the description of the first, the second, and the third embodiments, the master device is the
projector 100 a and the slave devices are theprojectors 100 b to 100 d. However, it is not limited to this. For example, the master device may be theprojector 100 d and the slave devices may be theprojectors 100 a to 100 c. - Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
- While the present disclosure has been described with reference to exemplary embodiments, it is to be understood that the disclosure is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
Claims (46)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/377,914 US20190235823A1 (en) | 2015-04-02 | 2019-04-08 | Projection device and control method therefor |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015-076125 | 2015-04-02 | ||
JP2015076125A JP6275075B2 (en) | 2015-04-02 | 2015-04-02 | Projection apparatus, control method, and program |
US15/084,325 US9875074B2 (en) | 2015-04-02 | 2016-03-29 | Projection device and control method therefor |
US15/844,352 US10394514B2 (en) | 2015-04-02 | 2017-12-15 | Projection device and control method therefor |
US16/377,914 US20190235823A1 (en) | 2015-04-02 | 2019-04-08 | Projection device and control method therefor |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/844,352 Division US10394514B2 (en) | 2015-04-02 | 2017-12-15 | Projection device and control method therefor |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190235823A1 true US20190235823A1 (en) | 2019-08-01 |
Family
ID=57015198
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/084,325 Active US9875074B2 (en) | 2015-04-02 | 2016-03-29 | Projection device and control method therefor |
US15/844,352 Active US10394514B2 (en) | 2015-04-02 | 2017-12-15 | Projection device and control method therefor |
US16/377,914 Abandoned US20190235823A1 (en) | 2015-04-02 | 2019-04-08 | Projection device and control method therefor |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/084,325 Active US9875074B2 (en) | 2015-04-02 | 2016-03-29 | Projection device and control method therefor |
US15/844,352 Active US10394514B2 (en) | 2015-04-02 | 2017-12-15 | Projection device and control method therefor |
Country Status (2)
Country | Link |
---|---|
US (3) | US9875074B2 (en) |
JP (1) | JP6275075B2 (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6798108B2 (en) * | 2016-01-20 | 2020-12-09 | セイコーエプソン株式会社 | Image projection system and control method of image projection system |
JP6707870B2 (en) * | 2016-01-20 | 2020-06-10 | セイコーエプソン株式会社 | Projection system and projection position detection method |
JP6946647B2 (en) * | 2017-01-11 | 2021-10-06 | セイコーエプソン株式会社 | Display device and control method |
JP7129149B2 (en) * | 2017-06-29 | 2022-09-01 | キヤノン株式会社 | Projector and its control method |
US11093197B2 (en) * | 2017-07-31 | 2021-08-17 | Stmicroelectronics, Inc. | System and method to increase display area utilizing a plurality of discrete displays |
CN107959836A (en) * | 2017-11-15 | 2018-04-24 | 苏州佳世达光电有限公司 | A kind of projecting method and optical projection system |
JP6770502B2 (en) | 2017-11-22 | 2020-10-14 | キヤノン株式会社 | Communication devices, display devices, their control methods, programs and display systems |
DE102018112873A1 (en) * | 2018-05-29 | 2019-12-05 | Ebm-Papst Mulfingen Gmbh & Co. Kg | Auto addressing with position determination of bus users |
CN109062531B (en) * | 2018-08-15 | 2021-08-27 | 京东方科技集团股份有限公司 | Spliced screen, display method of spliced screen and display control device |
JP7238298B2 (en) * | 2018-09-03 | 2023-03-14 | セイコーエプソン株式会社 | Projectors and projector control methods |
CN112328193A (en) * | 2020-09-09 | 2021-02-05 | 深圳Tcl新技术有限公司 | Multi-terminal screen combination method, device, equipment and computer storage medium |
CN113590067A (en) * | 2021-07-28 | 2021-11-02 | 深圳创维-Rgb电子有限公司 | Screen projection control method, system, device and computer readable storage medium |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2501965B2 (en) * | 1991-02-12 | 1996-05-29 | パイオニア株式会社 | Data transfer control device |
KR100465173B1 (en) * | 2002-11-18 | 2005-01-13 | 삼성전자주식회사 | Image displayer with separating screen function |
JP2006284990A (en) * | 2005-04-01 | 2006-10-19 | Matsushita Electric Ind Co Ltd | Image transmission multi-display system |
KR101234156B1 (en) * | 2006-06-14 | 2013-02-15 | 삼성전자주식회사 | Method for providing external-input list using item-grouping and video apparatus thereof |
JP2006338044A (en) * | 2006-07-24 | 2006-12-14 | Sony Corp | Multi-display device, and multi-display control method, and computer program |
US7901094B2 (en) * | 2007-01-24 | 2011-03-08 | Seiko Epson Corporation | View projection: one-touch setup of light displays on arbitrary surfaces |
JP5515351B2 (en) * | 2009-03-24 | 2014-06-11 | セイコーエプソン株式会社 | Image output apparatus, control method, and control program |
JP2010262201A (en) * | 2009-05-11 | 2010-11-18 | Seiko Epson Corp | Projector system, projector, method and program for controlling projector |
KR101694159B1 (en) * | 2010-04-21 | 2017-01-09 | 엘지전자 주식회사 | Augmented Remote Controller and Method of Operating the Same |
JP5641820B2 (en) * | 2010-08-25 | 2014-12-17 | キヤノン株式会社 | Projection type display system and projection type display device |
JP2014006357A (en) * | 2012-06-22 | 2014-01-16 | Seiko Epson Corp | Projector, image display system, and control method of the projector |
JP2014107713A (en) * | 2012-11-28 | 2014-06-09 | Seiko Epson Corp | Operation method, operation program and operation apparatus |
US9041863B2 (en) * | 2013-08-30 | 2015-05-26 | Xiaomi Inc. | Electronic device and method for displaying resources |
-
2015
- 2015-04-02 JP JP2015076125A patent/JP6275075B2/en active Active
-
2016
- 2016-03-29 US US15/084,325 patent/US9875074B2/en active Active
-
2017
- 2017-12-15 US US15/844,352 patent/US10394514B2/en active Active
-
2019
- 2019-04-08 US US16/377,914 patent/US20190235823A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
US20160291919A1 (en) | 2016-10-06 |
US9875074B2 (en) | 2018-01-23 |
JP6275075B2 (en) | 2018-02-07 |
US20180107443A1 (en) | 2018-04-19 |
JP2016197146A (en) | 2016-11-24 |
US10394514B2 (en) | 2019-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10394514B2 (en) | Projection device and control method therefor | |
US20170142382A1 (en) | Communication apparatus, method of controlling communication apparatus, non-transitory computer-readable storage medium | |
JP2014107713A (en) | Operation method, operation program and operation apparatus | |
JP5870586B2 (en) | Projector control device, display device, and program. | |
US10754237B2 (en) | Control apparatus, control method, projection system, and storage medium | |
US20160027402A1 (en) | Wireless communications system, and display apparatus | |
US10303419B2 (en) | Information processing system, display processing apparatus, display processing method, and recording medium | |
US10768884B2 (en) | Communication apparatus, display apparatus, control method thereof, storage medium, and display system for configuring multi-display settings | |
US20200169706A1 (en) | Projection control apparatus, control method of the same, and projection system | |
US20140211168A1 (en) | Image projection apparatus, control method, recording medium, and projection system | |
JP2015169940A (en) | Multi-projection system, information processor and program | |
US10009583B2 (en) | Projection system, projection apparatus, information processing method, and storage medium | |
US10484550B2 (en) | Display control apparatus that displays preview image, control method therefor, and storage medium | |
JP2017135490A (en) | Display system, display device, control device, and method for controlling these | |
JP2020178221A (en) | Projection control device, projection control method, and program | |
JP2006270461A (en) | Image formation apparatus and method | |
JP6825480B2 (en) | Information information system, display processing device, display processing method, and display program | |
US9881363B2 (en) | Image processing device, projector and image processing method for correcting a displayed image | |
US11206379B2 (en) | Controlled apparatus and control method thereof | |
JP7129149B2 (en) | Projector and its control method | |
JP2017073717A (en) | Information processing unit, information processing method, computer program and recording medium | |
US20190327457A1 (en) | Projection control apparatus and projection control method | |
JP6883196B2 (en) | Processing equipment, information processing equipment, processing systems, processing programs and information processing programs | |
CN115343898A (en) | Projection system and projection image superposition method | |
JP2012203311A (en) | Information processing system, information processing apparatus and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |