US20190235823A1 - Projection device and control method therefor - Google Patents

Projection device and control method therefor Download PDF

Info

Publication number
US20190235823A1
US20190235823A1 US16/377,914 US201916377914A US2019235823A1 US 20190235823 A1 US20190235823 A1 US 20190235823A1 US 201916377914 A US201916377914 A US 201916377914A US 2019235823 A1 US2019235823 A1 US 2019235823A1
Authority
US
United States
Prior art keywords
image
display apparatus
cpu
information
projector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/377,914
Inventor
Masaru Kurota
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to US16/377,914 priority Critical patent/US20190235823A1/en
Publication of US20190235823A1 publication Critical patent/US20190235823A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/1423Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
    • G06F3/1446Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display display composed of modules, e.g. video walls
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/64Constructional details of receivers, e.g. cabinets or dust covers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3141Constructional details thereof
    • H04N9/3147Multi-projection systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3179Video signal processing therefor
    • H04N9/3185Geometric adjustment, e.g. keystone or convergence
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0693Calibration of display systems
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2356/00Detection of the display position w.r.t. other display screens
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2370/00Aspects of data communication
    • G09G2370/20Details of the management of multiple sources of image data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/445Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information

Definitions

  • the present disclosure relates to a projection device and a control method therefor.
  • a multiple projection system which displays one projection image by synthesizing, on a screen, a plurality of images projected onto the screen by each of a plurality of projectors.
  • one projector (a master) of a plurality of projectors controls the other projectors (slaves).
  • Japanese Patent Application Laid-Open No. 2006-338044 discloses that, in the multiple projection system, one projector (a master) of a plurality of projectors associates identification information of the other projectors (slaves) with positional relationship of the other projectors (slaves). Further, it is known that the projector (master) notifies a position in one projection image displayed on a screen by the multiple projection system, to which a position of an image projected by another projector (slave) corresponds, to another projector (slave).
  • At least one aspect of the present disclosure is to solve all or at least one of the issues described above.
  • a projection device includes a projection unit configured to project an image, a communication unit configured to communicate with an external projection device, and an input unit configured to input an association between the external projection device and a predetermined area included in a display pattern for displaying one image on a screen by using an image projected by the projection device and an image projected by the external projection device.
  • the projection unit projects an image for notifying of the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit
  • the communication unit transmits information related to the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, to the external projection device.
  • FIG. 1 is a diagram illustrating an example of a multiple projection system in first to third embodiments.
  • FIGS. 2A and 2B are diagrams illustrating an example of a configuration of a projector in the first to the third embodiments.
  • FIG. 3 is a sequence diagram illustrating an example of projection processing performed in the first embodiment.
  • FIG. 4 is a diagram illustrating an example of a selection image in the first to the third embodiments.
  • FIGS. 5A and 5B are diagrams illustrating an example of an input image in the first to the third embodiments.
  • FIG. 6 is a diagram illustrating an example of an input image in the first to the third embodiments.
  • FIG. 7 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
  • FIGS. 8A and 8B are diagrams illustrating an example of an input image in the first to the third embodiments.
  • FIG. 9 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
  • FIG. 10 is a sequence diagram illustrating an example of projection processing in the second embodiment.
  • FIG. 11 is a diagram illustrating an example of an image projected by the multiple projection system in the second embodiment.
  • FIG. 12 is a flowchart illustrating an example of change processing in the third embodiment.
  • FIGS. 13A and 13B are diagrams illustrating an example of an image projected by the multiple projection system in the third embodiment.
  • FIG. 1 is a diagram illustrating a multiple projection system in the first embodiment.
  • the multiple projection system is a system that displays one image on a screen by synthesizing, on the screen, images projected onto the screen by each of a plurality of projectors.
  • the multiple projection system of the first embodiment includes four projectors 100 a to 100 d .
  • the projectors 100 a to 100 d respectively project four partial images 210 a to 210 d , which are obtained by dividing one image 200 to be displayed on a screen into four images, onto the screen.
  • the partial images 210 a to 210 d are displayed on the screen and one image corresponding to the image 200 is displayed on the screen by combining the four partial images 210 a to 210 d .
  • the partial image 210 a may include at least one of a part of the partial image 210 b , a part of the partial image 210 c , and a part of the partial image 210 d .
  • the partial image 210 b may include at least one of a part of the partial image 210 a , a part of the partial image 210 c , and a part of the partial image 210 d .
  • the partial image 210 c may include at least one of a part of the partial image 210 a , a part of the partial image 210 b , and a part of the partial image 210 d .
  • the partial image 210 d may include at least one of a part of the partial image 210 a , a part of the partial image 210 b , and a part of the partial image 210 c.
  • each of the projectors 100 a to 100 d projects a partial image obtained by dividing one image
  • an image projected onto the screen by each of the projectors 100 a to 100 d may be an image other than a partial image obtained by dividing one image.
  • the projector in the first embodiment may be rephrased to a “projection device” or may be rephrased to a “display device”.
  • the number of projectors included in the multiple projection system is not limited to four, but may be two.
  • the number of projectors included in the multiple projection system may be two or more.
  • the projector 100 a of the projectors 100 a to 100 d is set as a master device and the other projectors 100 b to 100 d are set as slave devices.
  • the projector 100 a transmits the image 200 to the projectors 100 b to 100 d .
  • each of the projectors 100 b to 100 d generates a partial image from the image 200 and projects the generated partial image onto the screen.
  • the projector 100 b generates a partial image 210 b from the image 200 received from the projector 100 a and projects the partial image 210 b onto the screen.
  • the projector 100 c generates a partial image 210 c from the image 200 received from the projector 100 a and projects the partial image 210 c onto the screen.
  • the projector 100 d generates a partial image 210 d from the image 200 received from the projector 100 a and projects the partial image 210 d onto the screen.
  • FIGS. 2A and 2B are diagrams illustrating an example of a configuration of the projector 100 a .
  • FIG. 2A is a diagram illustrating an example of a configuration of the projector 100 a and
  • FIG. 2B is an external view illustrating an example of the projector 100 a.
  • the projector 100 a includes a CPU (Central Processing Unit) 101 a , a ROM 102 a , a RAM 103 a , an image processing unit 104 a , a network IF (Interface) 105 a , and a user IF 106 a . Further, the projector 100 a includes an operation key 107 a , a remote control light-receiving unit 108 a , a projection unit 111 a , an input unit 112 a , and an internal bus 113 a.
  • the CPU 101 a is a control unit that controls each unit of the projector 100 a .
  • the ROM 102 a stores a computer program to be executed by the CPU 101 a .
  • the RAM 103 a temporarily stores a computer program and data as a work memory.
  • the CPU 101 a controls the projector 100 a by executing the program stored in the ROM 102 a.
  • the image processing unit 104 a performs image processing to generate an image to be projected by the projection unit 111 a .
  • the image processing unit 104 a performs light reduction processing (gradation processing) on an image that overlaps a part of an image projected by a projector other than the projector 100 a in an image projected by the projection unit 111 a .
  • the image that overlaps a part of an image projected by a projector other than the projector 100 a is referred to as an “overlapping area”.
  • the network IF 105 a communicates with an external projector (an external projection device) through a LAN (Local Area Network).
  • the network IF 105 a may communicate with an external projector through a wireless LAN or may communicate with an external projector through a wired LAN.
  • the network IF 105 a may communicate with an external projector based on a communication protocol other than LAN.
  • the user IF 106 a receives an input by a user from at least either one of the operation key 107 a and a remote control unit 109 a .
  • the remote control light-receiving unit 108 a supplies a signal received from the remote control unit 109 a to the user IF 106 a.
  • the projection unit 111 a projects an image supplied from the image processing unit 104 a .
  • the projection unit 111 a includes a liquid crystal panel and a driver that drives the liquid crystal panel. Further, the projection unit 111 a includes a drive system such as a lens and a zoom and a control system that controls a light source and a position of an image to be projected.
  • the input unit 112 a receives an image from outside.
  • the internal bus 113 a connects units included in the projector 100 a .
  • the remote control unit 109 a is used to control the projector 100 a.
  • the projectors 100 b to 100 d also have the same configuration as that of the projector 100 a .
  • the projector 100 b includes a CPU 101 b , a ROM 102 b , a RAM 103 b , an image processing unit 104 b , a network IF 105 b , and a user IF 106 b .
  • the projector 100 b includes an operation key 107 b , a remote control light-receiving unit 108 b , a projection unit 111 b , an input unit 112 b , and an internal bus 113 b .
  • the projector 100 c includes a CPU 101 c , a ROM 102 c , a RAM 103 c , an image processing unit 104 c , a network IF 105 c , and a user IF 106 c .
  • the projector 100 c includes an operation key 107 c , a remote control light-receiving unit 108 c , a projection unit 111 c , an input unit 112 c , and an internal bus 113 c .
  • the projector 100 d includes a CPU 101 d , a ROM 102 d , a RAM 103 d , an image processing unit 104 d , a network IF 105 d , and a user IF 106 d .
  • the projector 100 d includes an operation key 107 d , a remote control light-receiving unit 108 d , a projection unit 111 d , an input unit 112 d , and an internal bus 113 d.
  • the projector 100 a supplies an image received through at least either one of the input unit 112 a and the network IF 105 a to the image processing unit 104 a . Thereafter, the projector 100 a controls the projection unit 111 a so as to project an image on which image processing is performed by the image processing unit 104 a onto a screen. While the operation of the projector 100 a has been described, the projectors 100 b to 100 d also perform the same operation as that performed by the projector 100 a.
  • FIG. 3 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the first embodiment.
  • the projection processing illustrated in FIG. 3 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices.
  • the CPU 101 a When a user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the projector 100 a is selected as the master device, the CPU 101 a performs S 301 .
  • the CPU 101 a receives a master selection instruction through the user IF 106 a .
  • the CPU 101 a sets an operation mode of the projector 100 a to a master mode in which the projector 100 a operates as the master device. In this case, the CPU 101 a stores a value indicating that the projector 100 a is set to the master device into the RAM 103 a .
  • the CPU 101 a After S 301 is performed, the CPU 101 a performs S 302 .
  • the CPU 101 b When the user operates at least either one of the operation key 107 b and the remote control unit 109 b in the projector 100 b and thereby the projector 100 b is selected as the slave device, the CPU 101 b performs S 321 .
  • the CPU 101 b receives a slave selection instruction through the user IF 106 b .
  • the CPU 101 b sets an operation mode of the projector 100 b to a slave mode in which the projector 100 b operates as the slave device.
  • the CPU 101 b stores a value indicating that the projector 100 b is set to the slave device into the RAM 103 b .
  • the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 321 .
  • the CPU 101 a controls the projection unit 111 a so as to project a selection image 400 . Thereafter, the projection unit 111 a projects the selection image 400 onto a screen.
  • the selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system.
  • FIG. 4 is a diagram illustrating an example of the selection image 400 .
  • the selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system from among display patterns 401 to 404 .
  • the display patterns 401 to 404 are images which respectively represent display patterns different from each other.
  • the display pattern 401 is an image representing a pattern in which two partial images are arranged in the horizontal direction and displayed on the screen by two projectors. Further, in the display pattern 401 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the right partial image.
  • the display pattern 402 is an image representing a pattern of 2 ⁇ 2.
  • the display pattern 402 is an image representing a pattern in which four partial images are displayed on the screen by four projectors. Further, in the display pattern 402 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the upper left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the upper right partial image. Further, for example, identification information “3” is information for identifying a position of a predetermined area of the lower left partial image. Further, for example, identification information “4” is information for identifying a position of a predetermined area of the lower right partial image.
  • the display pattern 403 is an image representing a display pattern of 2 ⁇ 3.
  • the display pattern 403 is an image representing a pattern in which six partial images are displayed on the screen. Further, in the display pattern 403 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
  • the display pattern 404 is an image representing a display pattern of 3 ⁇ 3.
  • the display pattern 404 is an image representing a pattern in which nine partial images projected on the screen by nine projectors are displayed on the screen. Further, in the display pattern 403 , a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
  • the display patterns 401 to 404 and the selection image 400 are stored in advance in the ROM 102 a of the projector 100 a which is the master device.
  • the selection image 400 is displayed on the screen by the projection unit 111 a .
  • a user selects any one of the display patterns 401 to 404 by using at least either one of the operation key 107 a and the remote control unit 109 a .
  • the CPU 101 a performs S 303 .
  • the CPU 101 a receives a selection instruction of a display pattern through the user IF 106 a . Thereafter, the CPU 101 a selects any one of the display patterns 401 to 404 according to the selection instruction from the user and records information indicating the selected display pattern in the RAM 103 a . After S 303 is performed, the CPU 101 a performs S 304 .
  • the CPU 101 a selects the display pattern 402 and stores information indicating the display pattern 402 in the RAM 103 a.
  • the CPU 101 a controls the projection unit 111 a so as to project an input image 500 . Further, the CPU 101 a secures an area to assign IP addresses to a plurality of pieces of identification information of the display pattern selected in S 302 in the RAM 103 a.
  • the projection unit 111 a projects the input image 500 illustrated in FIG. 5A onto the screen in S 304 .
  • the input image 500 illustrated in FIG. 5A will be described.
  • the input image 500 illustrated in FIG. 5A includes an area 510 , an area 520 , and a confirmation button 530 .
  • the area 510 is an area where the display pattern 402 selected in S 302 is displayed.
  • the area 520 includes an area 521 where identification information for identifying positions of predetermined areas 510 a to 510 d included in the display pattern 402 is displayed and an input field 522 .
  • the display pattern 402 displayed in the area 510 includes the predetermined areas 510 a to 510 d , so that pieces of identification information “1” to “4” for identifying the positions of the predetermined areas 510 a to 510 d are displayed in the area 521 .
  • the input field 522 is an input field where IP addresses of the projectors are inputted with respect to the identification information “1” to “4” displayed in the area 521 .
  • the IP address is identification information for identifying each of the projectors 100 a to 100 d in a network for implementing the multiple projection system.
  • the area 521 includes areas 521 a to 521 d .
  • the input field 522 includes input fields 522 a to 522 d .
  • the identification information “1” for identifying the position of the predetermined area 510 a is displayed.
  • the input field 522 a is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 a .
  • the IP address inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”.
  • the input field 522 b is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 b .
  • the IP address inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”.
  • the input field 522 c is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 c .
  • the IP address inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”.
  • the input field 522 d is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 d .
  • the IP address inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”.
  • the input image 500 is used to cause the user to input IP addresses into the input fields 522 a to 522 d.
  • the input image 500 illustrated in FIG. 5A is displayed on the screen by the projection unit 111 a .
  • the user inputs IP addresses into the input field 522 as illustrated in FIG. 5B by using at least either one of the operation key 107 a and the remote control unit 109 a .
  • the IP address of the projector 100 a is “192.168.254.1” and the IP address of the projector 100 b is “192.168.254.2”.
  • the IP address of the projector 100 c is “192.168.254.3” and the IP address of the projector 100 d is “192.168.254.4”.
  • the user can input the IP addresses of the projectors into the input field 522 while confirming the positions where the projectors 100 a to 100 d are installed, the area 510 , and the area 521 .
  • the IP address of the projector 100 a which is the master device, is inputted into the input field 522 a in advance even when the user does not input the IP address.
  • the user should input the IP address of the projector 100 b into the input field 522 b . Further, the user should input the IP address of the projector 100 c into the input field 522 c . Further, the user should input the IP address of the projector 100 d into the input field 522 d.
  • the CPU 101 a monitors a subnet of the IP addresses inputted into the input field 522 .
  • the CPU 101 a controls the projection unit 111 a so as to project information indicating that there is an error in the input of the IP addresses onto the screen.
  • the information 600 such as “Subnets are different” as illustrated in FIG. 6 is projected by the projection unit 111 a .
  • the CPU 101 a performs S 305 .
  • the CPU 101 a sets projection positions of the projectors 100 a to 100 d according to the input into the input field 522 performed by the user.
  • the IP address “192.168.254.1” inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”.
  • the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.1” projects a partial image, to the position indicated by the identification information “1”. Therefore, the CPU 101 a stores first position information including the identification information “1” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.1”.
  • the first position information may further include information indicating the area 510 a.
  • the IP address “192.168.254.2” inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”.
  • the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.2” projects a partial image, to the position indicated by the identification information “2”. Therefore, the CPU 101 a stores second position information including the identification information “2” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.2”.
  • the second position information may further include information indicating the predetermined area 510 b.
  • the IP address “192.168.254.3” inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”.
  • the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.3” projects a partial image, to the position indicated by the identification information “3”. Therefore, the CPU 101 a stores third position information including the identification information “3” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.3”.
  • the third position information may further include information indicating the predetermined area 510 c.
  • the IP address “192.168.254.4” inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”.
  • the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.4” projects a partial image, to the position indicated by the identification information “4”. Therefore, the CPU 101 a stores fourth position information including the identification information “4” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.4”.
  • the fourth position information may further include information indicating the predetermined area 510 d.
  • the CPU 101 a When the user presses the confirmation button 530 by using at least either one of the operation key 107 a and the remote control unit 109 a , the CPU 101 a performs S 306 .
  • the confirmation button 530 When the confirmation button 530 is pressed, the assignment of the IP addresses by using the input image 500 is completed.
  • the CPU 101 a transmits the position information recorded in the RAM 103 a to the projectors 100 b to 100 d , which are slave devices, through the network IF 105 a.
  • the CPU 101 a transmits the first position information to the projector having the IP address “192.168.254.1”. Further, in S 306 , the CPU 101 a transmits the second position information to the projector having the IP address “192.168.254.2”. Further, in S 306 , the CPU 101 a transmits the third position information to the projector having the IP address “192.168.254.3”. Further, in S 306 , the CPU 101 a transmits the fourth position information to the projector having the IP address “192.168.254.4”. After all the position information recorded in the RAM 103 a is transmitted, the CPU 101 a performs S 307 .
  • the CPU 101 a controls the projection unit 111 a so as to project an image 700 a .
  • the image 700 a is an image for notifying of the display pattern selected in S 303 and the projection position of the projector 100 a.
  • the image 700 a in FIG. 7 is projected onto the screen by the projection unit 111 a .
  • the image 700 a includes an image representing the selected display pattern 402 and information (diagonal lines in the image 700 a ) indicating that the projection position of the projector 100 a is a position corresponding to the identification information “1”.
  • the projector 100 b When the projector 100 b receives the position information transmitted from the projector 100 a in S 306 , the projector 100 b performs S 322 .
  • the CPU 101 b controls the image processing unit 104 b so as to generate an image for notifying a position indicated by the position information received from the projector 100 a .
  • the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a.
  • the projector 100 b receives the second position information from the projector 100 a . Therefore, in S 322 , an image 700 b is projected on the screen by the projection unit 111 b .
  • the image 700 b is an image for notifying of a position indicated by the second position information.
  • the image 700 b includes information indicating the identification information “2” (“2” in the image 700 b ) included in the second position information.
  • the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b.
  • the projector 100 c receives the third position information from the projector 100 a . Therefore, in S 322 , an image 700 c is projected on the screen by the projection unit 111 c .
  • the image 700 c is an image for notifying of a position indicated by the third position information.
  • the image 700 c includes information indicating the identification information “3” (“3” in the image 700 c ) included in the third position information.
  • the projector 100 d receives the fourth position information from the projector 100 a . Therefore, in S 322 , an image 700 d is projected on the screen by the projection unit 111 d .
  • the image 700 d is an image for notifying of a position indicated by the fourth position information.
  • the image 700 d includes information indicating the identification information “4” (“4” in the image 700 d ) included in the fourth position information.
  • the images 700 a to 700 d are displayed on the screen as illustrated in FIG. 7 .
  • the user can identify the display pattern that is currently set and the projection position of the projector 100 a .
  • the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
  • the display pattern displayed in the image 700 a corresponds to the positions notified by the images 700 b to 700 d . Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
  • FIG. 8A illustrates a case in which the IP address of the projector 100 c is inputted in the input field 522 b and the IP address of the projector 100 b is inputted in the input field 522 c .
  • images 800 a to 800 d are displayed on the screen as illustrated in FIG. 8B .
  • the image 800 a is an image projected onto the screen by the projection unit 111 a in S 307 when the input of the IP addresses is as illustrated in FIG. 8A .
  • the image 800 a is the same image as the image 700 a .
  • the image 800 b is an image projected onto the screen by the projection unit 111 b in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
  • the image 800 c is an image projected onto the screen by the projection unit 111 c in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
  • the image 800 d is an image projected onto the screen by the projection unit 111 d in S 322 when the input of the IP addresses is as illustrated in FIG. 8A .
  • the user can identify the display pattern that is currently set and the projection position of the projector 100 a . Further, by checking the images 800 a to 800 d , the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
  • the images 800 a to 800 d as illustrated in FIG. 8B are displayed on the screen, the display pattern displayed in the image 800 a does not correspond to the positions notified by the images 800 b to 800 d . Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are not correctly associated with each other.
  • the user can recognize that there is an error in the input of the IP address of the projector 100 b and the IP address of the projector 100 c .
  • the user can correct the error in the input of the IP addresses by using at least either one of the operation key 107 a and the remote control unit 109 a.
  • a correction instruction to correct the input of the IP addresses is inputted into the CPU 101 a through the user IF 106 a .
  • the CPU 101 a determines whether or not the correction instruction is inputted through the user IF 106 a .
  • the CPU 101 a performs S 304 to S 307 again.
  • the CPU 101 a performs S 309 .
  • the CPU 101 a determines that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other and performs S 309 .
  • the user can issue a projection instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a .
  • the projection instruction is an instruction to start processing to display an image on the screen by the multiple projection system.
  • the CPU 101 a determines whether or not the projection instruction is inputted through the user IF 106 a .
  • the CPU 101 a performs S 310 .
  • the CPU 101 a returns to S 308 .
  • the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted into the input image 500 .
  • the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
  • the projector 100 b transmits the resolution information representing the resolution of the partial image projected by the projector 100 b to the projector 100 a through the network IF 105 b .
  • the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 323 .
  • the CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a .
  • the CPU 101 a requests the projectors 100 b to 100 d to transmit area information representing a position of an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a.
  • the overlapping area of the projector 100 b is a part of the image projected by the projector 100 b and the part overlaps with a part of images projected by projectors other than the projector 100 b .
  • the size and the position of the overlapping area of the projector 100 b is set by the user.
  • the light reduction processing is performed on the overlapping area of the projector 100 b by the image processing unit 104 b.
  • the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located upward.
  • the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located rightward.
  • the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located downward.
  • the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located leftward.
  • the CPU 101 b transmits the area information indicating the position of the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a .
  • the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 324 .
  • the CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
  • the CPU 101 a When all the resolution information and all the area information are acquired, the CPU 101 a performs S 311 .
  • the CPU 101 a calculates the size of the entire image displayed on the screen by using the resolution information of the projectors 100 a to 100 d and the area information of the projectors 100 a to 100 d .
  • the entire image is an image obtained by combining the partial image 210 a , the partial image 210 b , the partial image 210 c , and the partial image 210 d , which are illustrated in FIG. 1 .
  • the CPU 101 a calculates coordinates indicating an area of the partial image projected by each of the projectors 100 a to 100 d in the image 200 based on the size of the entire image, the ratio of the image 200 , and the display pattern selected in S 303 .
  • the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 a as first coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 b as second coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 c as third coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 d as fourth coordinate information. The CPU 101 a stores the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information, which are acquired, into the RAM 103 a . When the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information are calculated, the CPU 101 a performs S 312 .
  • the CPU 101 a transmits coordinate information and the image 200 to the projectors 100 b to 100 d through the network IF 105 a.
  • the CPU 101 a transmits the second coordinate information and the image 200 to the projector 100 b . Further, the CPU 101 a transmits the third coordinate information and the image 200 to the projector 100 c . Further, the CPU 101 a transmits the fourth coordinate information and the image 200 to the projector 100 d . After the second coordinate information, the third coordinate information, and the fourth coordinate information are transmitted, the CPU 101 a performs S 313 .
  • the CPU 101 a reads the first coordinate information from the RAM 103 a . Thereafter, the CPU 101 a controls the image processing unit 104 a so as to generate the partial image 210 a , which is an image to be projected by the projection unit 111 a , based on the image 200 and the first coordinate information. Further, the CPU 101 a controls the image processing unit 104 a so as to perform the light reduction processing on the overlapping area in the partial image 210 a generated by the image processing unit 104 a . Then, the CPU 101 a controls the projection unit 111 a so as to project the partial image 210 a on which the light reduction processing has been performed.
  • the projection unit 111 a projects the partial image 210 a , so that the partial image 210 a is displayed on the screen as illustrated in FIG. 1 .
  • the CPU 101 a performs S 314 .
  • the user can issue a projection end instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a .
  • the projection end instruction is an instruction to end the processing to display an image on the screen by the multiple projection system.
  • the CPU 101 a determines whether or not the projection end instruction is inputted.
  • the CPU 101 a ends the projection processing.
  • the CPU 101 a returns to S 312 and the projection unit 111 a continuously performs projection of the partial image 210 a.
  • the CPU 101 b controls the image processing unit 104 b so as to generate the partial image 210 b , which is an image to be projected by the projection unit 111 b , based on the image 200 and the second coordinate information. Further, the CPU 101 b controls the image processing unit 104 b so as to perform the light reduction processing on the overlapping area in the partial image 210 b generated by the image processing unit 104 b . Then, the CPU 101 b controls the projection unit 111 b so as to project the partial image 210 b on which the light reduction processing has been performed.
  • the projection unit 111 b projects the partial image 210 b , so that the partial image 210 b is displayed on the screen as illustrated in FIG. 1 .
  • the CPU 101 b performs S 326 .
  • the projectors 100 c and 100 d also perform, in S 325 , the same operation as that performed by the projector 100 b.
  • the user can issue the projection end instruction to the projector 100 b by using at least either one of the operation key 107 b and the remote control unit 109 b . Therefore, in S 326 , the CPU 101 b determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S 326 ), the CPU 101 b ends the projection processing. When the projection end instruction is not inputted (No in S 326 ), the CPU 101 b returns to S 325 and the projection unit 111 b continuously performs projection of the partial image 210 b.
  • the projectors 100 c and 100 d also perform, in S 326 , the same operation as that performed by the projector 100 b.
  • S 301 to S 314 are processing realized when the CPU 101 a of the projector 100 a , which is the master device, executes a computer program stored in the ROM 102 a .
  • S 321 to S 326 are processing realized when the CPU 101 b of the projector 100 b , which is the slave device, executes a computer program stored in the ROM 102 b .
  • S 321 to S 326 are processing realized when the CPU 101 c of the projector 100 c , which is the slave device, executes a computer program stored in the ROM 102 c .
  • S 321 to S 326 are processing realized when the CPU 101 d of the projector 100 d , which is the slave device, executes a computer program stored in the ROM 102 d.
  • the display patterns that can be selected by using the selection image 400 are the display patterns 401 to 404 .
  • the display patterns are not limited to these patterns.
  • the display patterns that can be selected by using the selection image 400 may further include display patterns other than the display patterns 401 to 404 .
  • the master device projects an image for notifying of the display pattern selected by the user and the projection position of the image to be projected by the master device.
  • the slave device projects an image for notifying of the projection position of the image to be projected by the slave device.
  • the user can confirm the selected display pattern and the projection positions of the images to be projected by the slave devices that are currently set. For example, when the positions indicated by the images projected by the slave devices do not correspond to the positions illustrated in the display pattern included in the image projected by the master device, the user can determine that there is an error in the input of the IP addresses of the slave devices. In this case, the user can easily determine that the positions associated with the IP addresses of the slave devices do not correspond to the current installation positions of the slave devices, so that the user can correctly modify the input of the IP addresses of the slave devices.
  • the multiple projection system of the first embodiment it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
  • the projector 100 a transmits position information to the projectors 100 b to 100 d and transmits information representing the display pattern selected in S 303 to the projectors 100 b to 100 d . Further, in S 307 , the projector 100 a controls the projection unit 111 a so as to project the image 900 a instead of the image 700 a . In S 322 , the projector 100 b controls the image processing unit 104 b so as to generate the image 900 b based on the position information and the information representing the display pattern which are received from the projector 100 a .
  • the projector 100 b controls the projection unit 111 b so as to project the image 900 b based on the position information received from the projector 100 a .
  • the projector 100 c controls the projection unit 111 c so as to project the generated image 900 c based on the position information and the information representing the display pattern which are received from the projector 100 a .
  • the projector 100 d controls the projection unit 111 d so as to project the generated image 900 d based on the position information and the information representing the display pattern which are received from the projector 100 a.
  • FIG. 10 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the second embodiment.
  • S 301 to S 305 , S 307 to S 314 , S 321 , S 325 , and S 326 in the projection processing in FIG. 10 are the same processing as S 301 to S 305 , S 307 to S 314 , S 321 , S 325 , and S 326 in the projection processing in FIG. 3 , so that the description thereof will be omitted.
  • the projection processing illustrated in FIG. 10 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices. Further, the projection processing illustrated in FIG. 10 will be described by using an example in which the display pattern 402 is selected in S 303 .
  • the CPU 101 a After performing S 305 , the CPU 101 a performs S 1001 .
  • the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted by the user in S 304 .
  • the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
  • the CPU 101 b transmits the resolution information representing the resolution of the image projected by the projector 100 b to the projector 100 a through the network IF 105 b .
  • the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 1011 .
  • the CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a .
  • the CPU 101 a requests the projectors 100 b to 100 d to transmit area information indicating an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a .
  • the CPU 101 b transmits the area information indicating the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a .
  • the projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S 1012 .
  • the CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
  • the CPU 101 a When all the resolution information and all the area information are acquired, the CPU 101 a performs S 1002 . In S 1002 , the CPU 101 a performs determination processing to determine whether or not the setting performed in S 305 is correct by using all the acquired area information.
  • the CPU 101 a performs first processing, second processing, third processing, and fourth processing.
  • the first processing is processing for determining whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not.
  • the CPU 101 a acquires first information indicating the position of the overlapping area included in the predetermined area 510 a from the RAM 103 a .
  • the first information is stored in the RAM 103 a in advance.
  • the right side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 b , so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located rightward.
  • the lower side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 c , so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located downward.
  • the CPU 101 a compares the first information with the area information acquired from the projector having the IP address inputted into the input field 522 a , and determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not according to the comparison result.
  • the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct.
  • the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is not correct.
  • the second processing is processing for determining whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not.
  • the CPU 101 a acquires second information indicating the position of the overlapping area included in the predetermined area 510 b from the RAM 103 a .
  • the second information is stored in the RAM 103 a in advance.
  • the left side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 a , so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located leftward.
  • the lower side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 d , so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located downward.
  • the CPU 101 a compares the second information with the area information acquired from the projector having the IP address inputted into the input field 522 b , and determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not according to the comparison result.
  • the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct.
  • the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct.
  • the third processing is processing for determining whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not.
  • the CPU 101 a acquires third information indicating the position of the overlapping area included in the predetermined area 510 c from the RAM 103 a .
  • the third information is stored in the RAM 103 a in advance.
  • the upper side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 a , so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located upward.
  • the right side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 d , so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located rightward.
  • the CPU 101 a compares the third information with the area information acquired from the projector having the IP address inputted into the input field 522 c , and determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not according to the comparison result.
  • the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct.
  • the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is not correct.
  • the fourth processing is processing for determining whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not.
  • the CPU 101 a acquires fourth information indicating the position of the overlapping area included in the predetermined area 510 d from the RAM 103 a .
  • the fourth information is stored in the RAM 103 a in advance.
  • the upper side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 b , so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located upward.
  • the left side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 c , so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located leftward.
  • the CPU 101 a compares the fourth information with the area information acquired from the projector having the IP address inputted into the input field 522 d , and determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not according to the comparison result.
  • the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct.
  • the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is not correct.
  • the CPU 101 a determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not. After S 1002 is performed, the CPU 101 a performs S 1003 .
  • the CPU 101 a transmits result information indicating the determination result of the determination processing in S 1002 and the position information recorded in the RAM 103 a to the projectors 100 b to 100 d through the network IF 105 a.
  • S 1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct will be described.
  • the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b .
  • the result information is information indicating that there is an error in the input of the IP address.
  • S 1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct will be described.
  • the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b .
  • the result information is information indicating that the input of the IP address is correct.
  • the CPU 101 a performs S 307 .
  • the projector 100 b When the projector 100 b receives the result information and the position information transmitted from the projector 100 a in S 1002 , the projector 100 b performs S 1013 .
  • the CPU 101 b controls the image processing unit 104 b so as to generate an image based on the position information and the result information received from the projector 100 a.
  • the image processing unit 104 b When the result information received from the projector 100 a indicates that there is an error in the input of the IP address, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information and for notifying that there is an error in the input of the IP address. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a . In this case, for example, as illustrated in FIG. 11 , the projector 100 b projects an image 1101 onto the screen. The image 1101 includes information 1102 indicating that there is an error in the input of the IP address.
  • the image processing unit 104 b When the result information received from the projector 100 a indicates that the input of the IP address is correct, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a .
  • the projector 100 b may control the projection unit 111 b so as to project an image indicating that the input of the IP address is correct. In this case, the projection unit 111 b projects an image indicating that the input of the IP address is correct at a position indicated by the position information received from the projector 100 a.
  • the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b .
  • the CPU 101 b performs S 325 .
  • S 1001 and S 1002 are processing realized when the CPU 101 a of the projector 100 a , which is the master device, executes a computer program stored in the ROM 102 a .
  • S 1011 to S 1013 are processing realized when the CPU 101 b of the projector 100 b , which is the slave device, executes a computer program stored in the ROM 102 b .
  • S 1011 to S 1013 are processing realized when the CPU 101 c of the projector 100 c , which is the slave device, executes a computer program stored in the ROM 102 c .
  • S 1011 to S 1013 are processing realized when the CPU 101 d of the projector 100 d , which is the slave device, executes a computer program stored in the ROM 102 d.
  • the master device projects an image for notifying of the display pattern selected by the user and the position of the image to be projected by the master device.
  • the slave device projects an image for notifying of the position of the image to be projected by the slave device.
  • the master device determines whether or not there is an error in the input of the IP address of the slave device based on the position of the overlapping area of the image projected by the slave device and notifies the slave device of the determination result. Further, when the slave device is notified by the master device that there is an error in the input of the IP address, the slave device projects an image for notifying information indicating that there is an error in the input of the IP address. Thereby, the user can easily determine that the position associated with the IP address of the slave device does not correspond to the current installation position of the slave device, so that the user can correctly modify the input of the IP address of the slave device.
  • the multiple projection system of the second embodiment it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
  • the projector 100 a notifies the projectors 100 b to 100 d of the determination result of the determination processing of S 1002 and the projectors 100 b to 100 d project the determination result of the determination processing of S 1002 which is notified from the master device.
  • the projector 100 a notifies the projectors 100 b to 100 d of the determination result of the determination processing of S 1002 and the projectors 100 b to 100 d project the determination result of the determination processing of S 1002 which is notified from the master device.
  • it is not limited to this.
  • the projector 100 a searches for information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b from among the first information, the second information, the third information, and the fourth information. Thereafter, the projector 100 a changes the setting of the IP address corresponding to the predetermined area 510 b according to the information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b .
  • the projector 100 a searches for information corresponding to the area information acquired from the projector 100 b from among the first information, the second information, the third information, and the fourth information.
  • the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct and then performs S 1003 .
  • the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct. Thereafter, the projector 100 a changes the setting of the IP address of the projector 100 b so that the IP address of the projector 100 b is inputted into the input field 522 c . Thereby, the predetermined area 510 c and the IP address of the projector 100 b inputted into the input field 522 c is associated with each other.
  • the projector 100 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct and performs S 1003 .
  • the projector 100 a may also perform the same processing for the other predetermined areas 510 a , 510 c , and 510 d.
  • a multiple projection system in the third embodiment will be described.
  • the description of the same configuration and processing as those in the first and the second embodiments will be omitted.
  • a multiple projection system will be described where when each of the projectors 100 a to 100 d projects a partial image based on, for example, the display pattern 402 , the display pattern is changed to a different display pattern by a user.
  • the CPU 101 a performs change processing illustrated in FIG. 12 .
  • the CPU 101 a When the user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the display pattern is changed, the CPU 101 a performs S 1201 .
  • the CPU 101 a receives a change instruction of the display pattern through the user IF 106 a . After S 1201 is performed, the CPU 101 a performs S 1202 . In S 1202 , the CPU 101 a updates the position information set in S 305 based on the display pattern after change indicated by the change instruction of the display pattern and the display pattern before change.
  • the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 401 .
  • the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 403 .
  • the identification information is “1” and “2” from the left (the first column) of the first row and “3” and “4” from the left (the first column) of the second row.
  • the identification information is “1”, “2”, and “3” from the left (the first column) of the first row and “4”, “5”, and “6” from the left (the first column) of the second row. Therefore, the CPU 101 a updates the position information stored in the RAM 103 a so that the display pattern after change corresponds to the installation positions of the projectors 100 a to 100 d.
  • the CPU 101 a associates the IP address associated with the identification information “3” (the second row, the first column) of the display pattern 402 with the identification information “4” (the first row, the third column) of the display pattern 403 and stores the IP address into the RAM 103 a . Further, the CPU 101 a associates the IP address associated with the identification information “4” (the second row, the second column) of the display pattern 402 with the identification information “5” (the second row, the first column) of the display pattern 403 and stores the IP address into the RAM 103 a .
  • the CPU 101 a connects to a projector through the network IF 105 a by using the IP address corresponding to the identification information “6”. If the CPU 101 a cannot establish a connection with the projector by using the IP address corresponding to the identification information “6”, it is assumed that no projector to be a connection destination is installed.
  • the CPU 101 a After S 1202 is performed, the CPU 101 a performs S 1203 .
  • the CPU 101 a requests resolution information from the projectors 100 b to 100 d .
  • the CPU 101 a acquires all the resolution information from the projectors 100 b to 100 d .
  • the CPU 101 a performs S 1204 .
  • the CPU 101 a determines the resolution of a projector of which resolution information cannot be acquired to be the same as the resolution of another projector. For example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b , the CPU 101 a determines that the resolution of the projector 100 b is the same as the resolution of the projector 100 c adjacent to the projector 100 b . Further, for example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b , the CPU 101 a may determine that the resolution of the projector 100 b is the same as the resolution of the projector 100 a which is the master device.
  • the CPU 101 a requests area information from the projectors 100 b to 100 d .
  • the CPU 101 a acquires all the area information from the projectors 100 b to 100 d .
  • the CPU 101 a performs S 1205 .
  • the CPU 101 a determines whether or not inconsistency occurs between adjacent sides based on the area information acquired in S 1204 .
  • the adjacent sides are, for example, the right side of the image displayed in the area corresponding to the identification information “1” and the left side of the image displayed in the area corresponding to the identification information “2” in the display pattern 402 .
  • the CPU 101 a determines that inconsistency occurs between adjacent sides (Yes in S 1205 ), the CPU 101 a performs S 1206 .
  • the CPU 101 a determines that no inconsistency occurs between adjacent sides (No in S 1205 )
  • the CPU 101 a performs S 1207 .
  • the CPU 101 a deletes the area information acquired in S 324 from the RAM 103 a and stores the area information acquired in S 1204 into the RAM 103 a .
  • the S 1206 is performed, the area information is updated.
  • the CPU 101 a does not perform the processing of S 1206 and maintains consistency between the projectors in the multiple projection system by using the area information acquired in S 324 .
  • FIG. 13A is a diagram illustrating an image 1301 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 401 .
  • FIG. 13B is a diagram illustrating an image 1302 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 403 .
  • the master device when the display pattern is changed by the user, the master device resets the projection positions of the slave devices based on the display pattern before change and the display pattern after change.
  • the multiple projection system in the third embodiment has the same effect as that of the first and the second embodiments regarding the same configuration and processing as those of the first and the second embodiments.
  • the master device is the projector 100 a and the slave devices are the projectors 100 b to 100 d .
  • the master device may be the projector 100 d and the slave devices may be the projectors 100 a to 100 c.
  • Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s).
  • computer executable instructions e.g., one or more programs
  • a storage medium which may also be referred to more fully as a
  • the computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions.
  • the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
  • the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Controls And Circuits For Display Device (AREA)
  • Transforming Electric Information Into Light Information (AREA)

Abstract

When each of a plurality of projection devices projects an image on a screen, in order to easily determine whether or not an association between identification information of each of the plurality of projection devices and positional relationship of each of the plurality of projection devices is correct, an image for notifying of an association between a predetermined area included in a display pattern and an external projection device is projected, and a communication unit of a projection device transmits information related to the association between the predetermined area included in the display pattern and the external projection device, which is inputted by an input unit, to the external projection device.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a divisional, and claims the benefit, of U.S. patent application Ser. No. 15/844,352, presently pending and filed on Dec. 15, 2017, which is a continuation, and claims the benefit, of U.S. patent application Ser. No. 15/084,325, filed on Mar. 29, 2016 and issued as U.S. Pat. No. 9,875,074 on Jan. 23, 2018, and claims the benefit of, and priority to, Japanese Patent Application No. 2015-076125, filed Apr. 2, 2015, which applications and patent(s) are hereby incorporated by reference herein in their entireties.
  • BACKGROUND OF THE INVENTION Field of the Invention
  • The present disclosure relates to a projection device and a control method therefor.
  • Description of the Related Art
  • Conventionally, a multiple projection system is known which displays one projection image by synthesizing, on a screen, a plurality of images projected onto the screen by each of a plurality of projectors.
  • It is known that, in the multiple projection system, one projector (a master) of a plurality of projectors controls the other projectors (slaves).
  • For example, Japanese Patent Application Laid-Open No. 2006-338044 discloses that, in the multiple projection system, one projector (a master) of a plurality of projectors associates identification information of the other projectors (slaves) with positional relationship of the other projectors (slaves). Further, it is known that the projector (master) notifies a position in one projection image displayed on a screen by the multiple projection system, to which a position of an image projected by another projector (slave) corresponds, to another projector (slave).
  • However, conventionally, a method that causes a user to easily determine whether the association between the identification information of another projector (slave) and the positional relationship of another projector (slave) is correct or not is not considered. Therefore, in a state in which the association between the identification information of another projector (slave) and the positional relationship of another projector (slave) is not correct, a user may cause all the projectors of the multiple projection system to project an image. Thereby, there is a possibility that a desired projection image is not displayed on a screen at a desired timing in the multiple projection system.
  • SUMMARY OF THE INVENTION
  • At least one aspect of the present disclosure is to solve all or at least one of the issues described above.
  • In an aspect of the present disclosure, a projection device includes a projection unit configured to project an image, a communication unit configured to communicate with an external projection device, and an input unit configured to input an association between the external projection device and a predetermined area included in a display pattern for displaying one image on a screen by using an image projected by the projection device and an image projected by the external projection device. In the projection device, the projection unit projects an image for notifying of the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, and the communication unit transmits information related to the association between the external projection device and the predetermined area included in the display pattern, which is inputted by the input unit, to the external projection device.
  • According to other aspects of the present disclosure, one or more additional projection devices, one or more control methods therefor and one or more mediums are discussed herein. Further features of the present disclosure will become apparent from the following description of exemplary embodiments (with reference to the attached drawings).
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate exemplary embodiments, features, and aspects of the disclosure and, together with the description, serve to explain the principles of the disclosure.
  • FIG. 1 is a diagram illustrating an example of a multiple projection system in first to third embodiments.
  • FIGS. 2A and 2B are diagrams illustrating an example of a configuration of a projector in the first to the third embodiments.
  • FIG. 3 is a sequence diagram illustrating an example of projection processing performed in the first embodiment.
  • FIG. 4 is a diagram illustrating an example of a selection image in the first to the third embodiments.
  • FIGS. 5A and 5B are diagrams illustrating an example of an input image in the first to the third embodiments.
  • FIG. 6 is a diagram illustrating an example of an input image in the first to the third embodiments.
  • FIG. 7 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
  • FIGS. 8A and 8B are diagrams illustrating an example of an input image in the first to the third embodiments.
  • FIG. 9 is a diagram illustrating an example of an image projected by the multiple projection system in the first embodiment.
  • FIG. 10 is a sequence diagram illustrating an example of projection processing in the second embodiment.
  • FIG. 11 is a diagram illustrating an example of an image projected by the multiple projection system in the second embodiment.
  • FIG. 12 is a flowchart illustrating an example of change processing in the third embodiment.
  • FIGS. 13A and 13B are diagrams illustrating an example of an image projected by the multiple projection system in the third embodiment.
  • DESCRIPTION OF THE EMBODIMENTS
  • Various exemplary embodiments, features, and aspects of the disclosure will be described in detail below with reference to the drawings.
  • First Embodiment
  • FIG. 1 is a diagram illustrating a multiple projection system in the first embodiment. The multiple projection system is a system that displays one image on a screen by synthesizing, on the screen, images projected onto the screen by each of a plurality of projectors.
  • The multiple projection system of the first embodiment includes four projectors 100 a to 100 d. The projectors 100 a to 100 d respectively project four partial images 210 a to 210 d, which are obtained by dividing one image 200 to be displayed on a screen into four images, onto the screen. Thereby, the partial images 210 a to 210 d are displayed on the screen and one image corresponding to the image 200 is displayed on the screen by combining the four partial images 210 a to 210 d. The partial image 210 a may include at least one of a part of the partial image 210 b, a part of the partial image 210 c, and a part of the partial image 210 d. In the same manner, the partial image 210 b may include at least one of a part of the partial image 210 a, a part of the partial image 210 c, and a part of the partial image 210 d. In the same manner, the partial image 210 c may include at least one of a part of the partial image 210 a, a part of the partial image 210 b, and a part of the partial image 210 d. In the same manner, the partial image 210 d may include at least one of a part of the partial image 210 a, a part of the partial image 210 b, and a part of the partial image 210 c.
  • In the first embodiment, a case where each of the projectors 100 a to 100 d projects a partial image obtained by dividing one image is described as an example. However, an image projected onto the screen by each of the projectors 100 a to 100 d may be an image other than a partial image obtained by dividing one image. The projector in the first embodiment may be rephrased to a “projection device” or may be rephrased to a “display device”.
  • The number of projectors included in the multiple projection system is not limited to four, but may be two. The number of projectors included in the multiple projection system may be two or more.
  • In the description of the first embodiment, it is assumed that the projector 100 a of the projectors 100 a to 100 d is set as a master device and the other projectors 100 b to 100 d are set as slave devices. In the multiple projection system, the projector 100 a transmits the image 200 to the projectors 100 b to 100 d. Further, each of the projectors 100 b to 100 d generates a partial image from the image 200 and projects the generated partial image onto the screen. For example, the projector 100 b generates a partial image 210 b from the image 200 received from the projector 100 a and projects the partial image 210 b onto the screen. Further, for example, the projector 100 c generates a partial image 210 c from the image 200 received from the projector 100 a and projects the partial image 210 c onto the screen. Further, for example, the projector 100 d generates a partial image 210 d from the image 200 received from the projector 100 a and projects the partial image 210 d onto the screen.
  • FIGS. 2A and 2B are diagrams illustrating an example of a configuration of the projector 100 a. FIG. 2A is a diagram illustrating an example of a configuration of the projector 100 a and FIG. 2B is an external view illustrating an example of the projector 100 a.
  • The projector 100 a includes a CPU (Central Processing Unit) 101 a, a ROM 102 a, a RAM 103 a, an image processing unit 104 a, a network IF (Interface) 105 a, and a user IF 106 a. Further, the projector 100 a includes an operation key 107 a, a remote control light-receiving unit 108 a, a projection unit 111 a, an input unit 112 a, and an internal bus 113 a.
  • The CPU 101 a is a control unit that controls each unit of the projector 100 a. The ROM 102 a stores a computer program to be executed by the CPU 101 a. The RAM 103 a temporarily stores a computer program and data as a work memory. The CPU 101 a controls the projector 100 a by executing the program stored in the ROM 102 a.
  • The image processing unit 104 a performs image processing to generate an image to be projected by the projection unit 111 a. For example, the image processing unit 104 a performs light reduction processing (gradation processing) on an image that overlaps a part of an image projected by a projector other than the projector 100 a in an image projected by the projection unit 111 a. The image that overlaps a part of an image projected by a projector other than the projector 100 a is referred to as an “overlapping area”.
  • The network IF 105 a communicates with an external projector (an external projection device) through a LAN (Local Area Network). The network IF 105 a may communicate with an external projector through a wireless LAN or may communicate with an external projector through a wired LAN. The network IF 105 a may communicate with an external projector based on a communication protocol other than LAN.
  • The user IF 106 a receives an input by a user from at least either one of the operation key 107 a and a remote control unit 109 a. The remote control light-receiving unit 108 a supplies a signal received from the remote control unit 109 a to the user IF 106 a.
  • The projection unit 111 a projects an image supplied from the image processing unit 104 a. The projection unit 111 a includes a liquid crystal panel and a driver that drives the liquid crystal panel. Further, the projection unit 111 a includes a drive system such as a lens and a zoom and a control system that controls a light source and a position of an image to be projected. The input unit 112 a receives an image from outside. The internal bus 113 a connects units included in the projector 100 a. The remote control unit 109 a is used to control the projector 100 a.
  • While the projector 100 a has been described, the projectors 100 b to 100 d also have the same configuration as that of the projector 100 a. For example, in the same manner as the projector 100 a, the projector 100 b includes a CPU 101 b, a ROM 102 b, a RAM 103 b, an image processing unit 104 b, a network IF 105 b, and a user IF 106 b. Further, the projector 100 b includes an operation key 107 b, a remote control light-receiving unit 108 b, a projection unit 111 b, an input unit 112 b, and an internal bus 113 b. For example, in the same manner as the projector 100 a, the projector 100 c includes a CPU 101 c, a ROM 102 c, a RAM 103 c, an image processing unit 104 c, a network IF 105 c, and a user IF 106 c. Further, the projector 100 c includes an operation key 107 c, a remote control light-receiving unit 108 c, a projection unit 111 c, an input unit 112 c, and an internal bus 113 c. For example, in the same manner as the projector 100 a, the projector 100 d includes a CPU 101 d, a ROM 102 d, a RAM 103 d, an image processing unit 104 d, a network IF 105 d, and a user IF 106 d. Further, the projector 100 d includes an operation key 107 d, a remote control light-receiving unit 108 d, a projection unit 111 d, an input unit 112 d, and an internal bus 113 d.
  • Next, an operation of the projector 100 a will be described. The projector 100 a supplies an image received through at least either one of the input unit 112 a and the network IF 105 a to the image processing unit 104 a. Thereafter, the projector 100 a controls the projection unit 111 a so as to project an image on which image processing is performed by the image processing unit 104 a onto a screen. While the operation of the projector 100 a has been described, the projectors 100 b to 100 d also perform the same operation as that performed by the projector 100 a.
  • FIG. 3 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the first embodiment. The projection processing illustrated in FIG. 3 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices.
  • When a user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the projector 100 a is selected as the master device, the CPU 101 a performs S301. In S301, the CPU 101 a receives a master selection instruction through the user IF 106 a. Thereafter, the CPU 101 a sets an operation mode of the projector 100 a to a master mode in which the projector 100 a operates as the master device. In this case, the CPU 101 a stores a value indicating that the projector 100 a is set to the master device into the RAM 103 a. After S301 is performed, the CPU 101 a performs S302.
  • When the user operates at least either one of the operation key 107 b and the remote control unit 109 b in the projector 100 b and thereby the projector 100 b is selected as the slave device, the CPU 101 b performs S321. In S321, the CPU 101 b receives a slave selection instruction through the user IF 106 b. Thereafter, the CPU 101 b sets an operation mode of the projector 100 b to a slave mode in which the projector 100 b operates as the slave device. In this case, the CPU 101 b stores a value indicating that the projector 100 b is set to the slave device into the RAM 103 b. The projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S321.
  • In S302, the CPU 101 a controls the projection unit 111 a so as to project a selection image 400. Thereafter, the projection unit 111 a projects the selection image 400 onto a screen. The selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system.
  • FIG. 4 is a diagram illustrating an example of the selection image 400. The selection image 400 is an image to select a display pattern of an image to be displayed on the screen by the multiple projection system from among display patterns 401 to 404. The display patterns 401 to 404 are images which respectively represent display patterns different from each other.
  • The display pattern 401 is an image representing a pattern in which two partial images are arranged in the horizontal direction and displayed on the screen by two projectors. Further, in the display pattern 401, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the right partial image.
  • The display pattern 402 is an image representing a pattern of 2×2. The display pattern 402 is an image representing a pattern in which four partial images are displayed on the screen by four projectors. Further, in the display pattern 402, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other. For example, identification information “1” is information for identifying a position of a predetermined area of the upper left partial image. Further, for example, identification information “2” is information for identifying a position of a predetermined area of the upper right partial image. Further, for example, identification information “3” is information for identifying a position of a predetermined area of the lower left partial image. Further, for example, identification information “4” is information for identifying a position of a predetermined area of the lower right partial image.
  • The display pattern 403 is an image representing a display pattern of 2×3. The display pattern 403 is an image representing a pattern in which six partial images are displayed on the screen. Further, in the display pattern 403, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
  • The display pattern 404 is an image representing a display pattern of 3×3. The display pattern 404 is an image representing a pattern in which nine partial images projected on the screen by nine projectors are displayed on the screen. Further, in the display pattern 403, a predetermined area which is an area where one partial image is projected and identification information for identifying a position of the predetermined area are associated with each other.
  • The display patterns 401 to 404 and the selection image 400 are stored in advance in the ROM 102 a of the projector 100 a which is the master device.
  • In S302, the selection image 400 is displayed on the screen by the projection unit 111 a. In this case, a user selects any one of the display patterns 401 to 404 by using at least either one of the operation key 107 a and the remote control unit 109 a. When the user selects any one of the display patterns 401 to 404 by using at least either one of the operation key 107 a and the remote control unit 109 a, the CPU 101 a performs S303.
  • In S303, the CPU 101 a receives a selection instruction of a display pattern through the user IF 106 a. Thereafter, the CPU 101 a selects any one of the display patterns 401 to 404 according to the selection instruction from the user and records information indicating the selected display pattern in the RAM 103 a. After S303 is performed, the CPU 101 a performs S304.
  • Hereinafter, a case in which the display pattern 402 is selected by the user in S303 will be described as an example. In this case, in S303, the CPU 101 a selects the display pattern 402 and stores information indicating the display pattern 402 in the RAM 103 a.
  • In S304, the CPU 101 a controls the projection unit 111 a so as to project an input image 500. Further, the CPU 101 a secures an area to assign IP addresses to a plurality of pieces of identification information of the display pattern selected in S302 in the RAM 103 a.
  • For example, when the display pattern 402 is selected in S303, the projection unit 111 a projects the input image 500 illustrated in FIG. 5A onto the screen in S304.
  • The input image 500 illustrated in FIG. 5A will be described. The input image 500 illustrated in FIG. 5A includes an area 510, an area 520, and a confirmation button 530. The area 510 is an area where the display pattern 402 selected in S302 is displayed. The area 520 includes an area 521 where identification information for identifying positions of predetermined areas 510 a to 510 d included in the display pattern 402 is displayed and an input field 522. The display pattern 402 displayed in the area 510 includes the predetermined areas 510 a to 510 d, so that pieces of identification information “1” to “4” for identifying the positions of the predetermined areas 510 a to 510 d are displayed in the area 521. The input field 522 is an input field where IP addresses of the projectors are inputted with respect to the identification information “1” to “4” displayed in the area 521. The IP address is identification information for identifying each of the projectors 100 a to 100 d in a network for implementing the multiple projection system.
  • Next, the area 521 and the input field 522 will be described with reference to FIG. 5B. The area 521 includes areas 521 a to 521 d. The input field 522 includes input fields 522 a to 522 d. In the area 521 a, the identification information “1” for identifying the position of the predetermined area 510 a is displayed. The input field 522 a is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 a. The IP address inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”.
  • In the area 521 b, the identification information “2” for identifying the position of the predetermined area 510 b is displayed. The input field 522 b is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 b. The IP address inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”.
  • In the area 521 c, the identification information “3” for identifying the position of the predetermined area 510 c is displayed. The input field 522 c is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 c. The IP address inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”.
  • In the area 521 d, the identification information “4” for identifying the position of the predetermined area 510 d is displayed. The input field 522 d is an input field into which the user is caused to input an IP address of a projector that projects an image to the position of the predetermined area 510 d. The IP address inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”.
  • To display an image on the screen according to the display pattern 402, it is necessary to assign IP addresses of projectors to the predetermined areas 510 a to 510 d. Therefore, the input image 500 is used to cause the user to input IP addresses into the input fields 522 a to 522 d.
  • In S304, the input image 500 illustrated in FIG. 5A is displayed on the screen by the projection unit 111 a. Thereafter, the user inputs IP addresses into the input field 522 as illustrated in FIG. 5B by using at least either one of the operation key 107 a and the remote control unit 109 a. In the first embodiment, the IP address of the projector 100 a is “192.168.254.1” and the IP address of the projector 100 b is “192.168.254.2”. Further, in the first embodiment, the IP address of the projector 100 c is “192.168.254.3” and the IP address of the projector 100 d is “192.168.254.4”.
  • In a state in which the input image 500 illustrated in FIG. 5A is displayed on the screen, the user can input the IP addresses of the projectors into the input field 522 while confirming the positions where the projectors 100 a to 100 d are installed, the area 510, and the area 521. In the first embodiment, it is assumed that the IP address of the projector 100 a, which is the master device, is inputted into the input field 522 a in advance even when the user does not input the IP address.
  • For example, a case in which the state where the projectors 100 a to 100 d are installed as illustrated in FIG. 1 will be described. In this case, the user should input the IP address of the projector 100 b into the input field 522 b. Further, the user should input the IP address of the projector 100 c into the input field 522 c. Further, the user should input the IP address of the projector 100 d into the input field 522 d.
  • Further, the CPU 101 a monitors a subnet of the IP addresses inputted into the input field 522. When the CPU 101 a receives input of a plurality of IP addresses of which subnets are different through the user IF 106 a, the CPU 101 a controls the projection unit 111 a so as to project information indicating that there is an error in the input of the IP addresses onto the screen. In this case, the information 600 such as “Subnets are different” as illustrated in FIG. 6 is projected by the projection unit 111 a. Thereby, the user can recognize that there is an error in the input of the IP addresses. When the user inputs the IP addresses, the CPU 101 a performs S305.
  • In S305, the CPU 101 a sets projection positions of the projectors 100 a to 100 d according to the input into the input field 522 performed by the user.
  • For example, in the case of FIG. 5B, in S305, the IP address “192.168.254.1” inputted into the input field 522 a is associated with the predetermined area 510 a and the identification information “1”. In this case, the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.1” projects a partial image, to the position indicated by the identification information “1”. Therefore, the CPU 101 a stores first position information including the identification information “1” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.1”. The first position information may further include information indicating the area 510 a.
  • For example, in the case of FIG. 5B, in S305, the IP address “192.168.254.2” inputted into the input field 522 b is associated with the predetermined area 510 b and the identification information “2”. In this case, the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.2” projects a partial image, to the position indicated by the identification information “2”. Therefore, the CPU 101 a stores second position information including the identification information “2” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.2”. The second position information may further include information indicating the predetermined area 510 b.
  • For example, in the case of FIG. 5B, in S305, the IP address “192.168.254.3” inputted into the input field 522 c is associated with the predetermined area 510 c and the identification information “3”. In this case, the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.3” projects a partial image, to the position indicated by the identification information “3”. Therefore, the CPU 101 a stores third position information including the identification information “3” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.3”. The third position information may further include information indicating the predetermined area 510 c.
  • For example, in the case of FIG. 5B, in S305, the IP address “192.168.254.4” inputted into the input field 522 d is associated with the predetermined area 510 d and the identification information “4”. In this case, the CPU 101 a sets the projection position, where the projector having the IP address “192.168.254.4” projects a partial image, to the position indicated by the identification information “4”. Therefore, the CPU 101 a stores fourth position information including the identification information “4” into the RAM 103 a as information indicating the projection position of the projector having the IP address “192.168.254.4”. The fourth position information may further include information indicating the predetermined area 510 d.
  • When the user presses the confirmation button 530 by using at least either one of the operation key 107 a and the remote control unit 109 a, the CPU 101 a performs S306. When the confirmation button 530 is pressed, the assignment of the IP addresses by using the input image 500 is completed.
  • In S306, the CPU 101 a transmits the position information recorded in the RAM 103 a to the projectors 100 b to 100 d, which are slave devices, through the network IF 105 a.
  • For example, in S306, the CPU 101 a transmits the first position information to the projector having the IP address “192.168.254.1”. Further, in S306, the CPU 101 a transmits the second position information to the projector having the IP address “192.168.254.2”. Further, in S306, the CPU 101 a transmits the third position information to the projector having the IP address “192.168.254.3”. Further, in S306, the CPU 101 a transmits the fourth position information to the projector having the IP address “192.168.254.4”. After all the position information recorded in the RAM 103 a is transmitted, the CPU 101 a performs S307.
  • In S307, the CPU 101 a controls the projection unit 111 a so as to project an image 700 a. The image 700 a is an image for notifying of the display pattern selected in S303 and the projection position of the projector 100 a.
  • When the display pattern 402 is selected in S303, the image 700 a in FIG. 7 is projected onto the screen by the projection unit 111 a. The image 700 a includes an image representing the selected display pattern 402 and information (diagonal lines in the image 700 a) indicating that the projection position of the projector 100 a is a position corresponding to the identification information “1”.
  • When the projector 100 b receives the position information transmitted from the projector 100 a in S306, the projector 100 b performs S322. In S322, the CPU 101 b controls the image processing unit 104 b so as to generate an image for notifying a position indicated by the position information received from the projector 100 a. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a.
  • For example, when the IP address of the projector 100 b is inputted into the input field 522 b by the user, the projector 100 b receives the second position information from the projector 100 a. Therefore, in S322, an image 700 b is projected on the screen by the projection unit 111 b. The image 700 b is an image for notifying of a position indicated by the second position information. The image 700 b includes information indicating the identification information “2” (“2” in the image 700 b) included in the second position information.
  • While the operation in which the projector 100 b performs S322 has been described, the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b.
  • For example, when the IP address of the projector 100 c is inputted into the input field 522 c by the user, the projector 100 c receives the third position information from the projector 100 a. Therefore, in S322, an image 700 c is projected on the screen by the projection unit 111 c. The image 700 c is an image for notifying of a position indicated by the third position information. The image 700 c includes information indicating the identification information “3” (“3” in the image 700 c) included in the third position information.
  • For example, when the IP address of the projector 100 d is inputted into the input field 522 d by the user, the projector 100 d receives the fourth position information from the projector 100 a. Therefore, in S322, an image 700 d is projected on the screen by the projection unit 111 d. The image 700 d is an image for notifying of a position indicated by the fourth position information. The image 700 d includes information indicating the identification information “4” (“4” in the image 700 d) included in the fourth position information.
  • When S307 is performed by the projector 100 a and S322 is performed by the projectors 100 b to 100 d, the images 700 a to 700 d are displayed on the screen as illustrated in FIG. 7. In this case, by checking the image 700 a, the user can identify the display pattern that is currently set and the projection position of the projector 100 a. Further, by checking the images 700 a to 700 d, the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other. When the images 700 b to 700 d as illustrated in FIG. 7 are displayed on the screen, the display pattern displayed in the image 700 a corresponds to the positions notified by the images 700 b to 700 d. Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other.
  • When all the IP addresses of the projectors 100 b to 100 d are correctly inputted in the input field 522 as illustrated in FIG. 5B, the images 700 b to 700 d as illustrated in FIG. 7 are displayed on the screen by performing S307 and S322.
  • However, there is a case in which the IP addresses of the projectors 100 b to 100 d are not correctly inputted in the input field 522. This case will be described with reference to FIG. 8A. FIG. 8A illustrates a case in which the IP address of the projector 100 c is inputted in the input field 522 b and the IP address of the projector 100 b is inputted in the input field 522 c. In this case, it is difficult for the user to recognize that there is an error in the input of the IP addresses of the projector 100 b and the projector 100 c. However, in this case, when S307 is performed by the projector 100 a and S322 is performed by the projectors 100 b to 100 d, images 800 a to 800 d are displayed on the screen as illustrated in FIG. 8B. The image 800 a is an image projected onto the screen by the projection unit 111 a in S307 when the input of the IP addresses is as illustrated in FIG. 8A. The image 800 a is the same image as the image 700 a. The image 800 b is an image projected onto the screen by the projection unit 111 b in S322 when the input of the IP addresses is as illustrated in FIG. 8A. The image 800 c is an image projected onto the screen by the projection unit 111 c in S322 when the input of the IP addresses is as illustrated in FIG. 8A. The image 800 d is an image projected onto the screen by the projection unit 111 d in S322 when the input of the IP addresses is as illustrated in FIG. 8A.
  • In this case, by checking the image 800 a, the user can identify the display pattern that is currently set and the projection position of the projector 100 a. Further, by checking the images 800 a to 800 d, the user can compare the display pattern that is currently set with the projection positions of the projectors 100 b to 100 d that are currently installed. Thereby, the user can easily determine whether or not the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other. When the images 800 a to 800 d as illustrated in FIG. 8B are displayed on the screen, the display pattern displayed in the image 800 a does not correspond to the positions notified by the images 800 b to 800 d. Therefore, the user can determine that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are not correctly associated with each other.
  • When the images 800 a to 800 d as illustrated in FIG. 8B are displayed, the user can recognize that there is an error in the input of the IP address of the projector 100 b and the IP address of the projector 100 c. In this case, the user can correct the error in the input of the IP addresses by using at least either one of the operation key 107 a and the remote control unit 109 a.
  • When the user performs an operation to correct the input of the IP addresses by using at least either one of the operation key 107 a and the remote control unit 109 a, a correction instruction to correct the input of the IP addresses is inputted into the CPU 101 a through the user IF 106 a. In S308, the CPU 101 a determines whether or not the correction instruction is inputted through the user IF 106 a. When the correction instruction is inputted (Yes in S308), the CPU 101 a performs S304 to S307 again. When the correction instruction is not inputted (No in S308), the CPU 101 a performs S309.
  • When the correction instruction is not inputted (No in S308), the CPU 101 a determines that the IP addresses of the projectors 100 b to 100 d and the positional relationship between the projectors 100 b to 100 d are correctly associated with each other and performs S309. When the correction instruction is not inputted (No in S308), the user can issue a projection instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a. The projection instruction is an instruction to start processing to display an image on the screen by the multiple projection system.
  • In S309, the CPU 101 a determines whether or not the projection instruction is inputted through the user IF 106 a. When the projection instruction is inputted (Yes in S309), the CPU 101 a performs S310. When the projection instruction is not inputted (No in S309), the CPU 101 a returns to S308.
  • In S310, the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted into the input image 500.
  • After S310 is performed, the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a. Thereafter, in S323, the projector 100 b transmits the resolution information representing the resolution of the partial image projected by the projector 100 b to the projector 100 a through the network IF 105 b. The projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S323. The CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a. Thereafter, the CPU 101 a requests the projectors 100 b to 100 d to transmit area information representing a position of an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a.
  • The overlapping area of the projector 100 b is a part of the image projected by the projector 100 b and the part overlaps with a part of images projected by projectors other than the projector 100 b. The size and the position of the overlapping area of the projector 100 b is set by the user. The light reduction processing is performed on the overlapping area of the projector 100 b by the image processing unit 104 b.
  • When the upper side of the image projected by the projector 100 b overlaps with a part of an image projected by projectors other than the projector 100 b, the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located upward. When the right side of the image projected by the projector 100 b overlaps with a part of an image projected by projectors other than the projector 100 b, the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located rightward. When the lower side of the image projected by the projector 100 b overlaps with a part of an image projected by projectors other than the projector 100 b, the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located downward. When the left side of the image projected by the projector 100 b overlaps with a part of an image projected by projectors other than the projector 100 b, the area information of the projector 100 b includes information indicating that the overlapping area of the projector 100 b is located leftward.
  • In S324, the CPU 101 b transmits the area information indicating the position of the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a. The projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S324. The CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
  • It is assumed that the resolution information of the projector 100 a and the area information of the projector 100 a are stored in the RAM 103 a.
  • When all the resolution information and all the area information are acquired, the CPU 101 a performs S311. In S311, the CPU 101 a calculates the size of the entire image displayed on the screen by using the resolution information of the projectors 100 a to 100 d and the area information of the projectors 100 a to 100 d. The entire image is an image obtained by combining the partial image 210 a, the partial image 210 b, the partial image 210 c, and the partial image 210 d, which are illustrated in FIG. 1. Further, the CPU 101 a calculates coordinates indicating an area of the partial image projected by each of the projectors 100 a to 100 d in the image 200 based on the size of the entire image, the ratio of the image 200, and the display pattern selected in S303.
  • For example, in S311, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 a as first coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 b as second coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 c as third coordinate information. Further, the CPU 101 a calculates coordinates indicating the area of the partial image of the projector 100 d as fourth coordinate information. The CPU 101 a stores the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information, which are acquired, into the RAM 103 a. When the first coordinate information, the second coordinate information, the third coordinate information, and the fourth coordinate information are calculated, the CPU 101 a performs S312.
  • In S312, the CPU 101 a transmits coordinate information and the image 200 to the projectors 100 b to 100 d through the network IF 105 a.
  • For example, in S312, the CPU 101 a transmits the second coordinate information and the image 200 to the projector 100 b. Further, the CPU 101 a transmits the third coordinate information and the image 200 to the projector 100 c. Further, the CPU 101 a transmits the fourth coordinate information and the image 200 to the projector 100 d. After the second coordinate information, the third coordinate information, and the fourth coordinate information are transmitted, the CPU 101 a performs S313.
  • In S313, the CPU 101 a reads the first coordinate information from the RAM 103 a. Thereafter, the CPU 101 a controls the image processing unit 104 a so as to generate the partial image 210 a, which is an image to be projected by the projection unit 111 a, based on the image 200 and the first coordinate information. Further, the CPU 101 a controls the image processing unit 104 a so as to perform the light reduction processing on the overlapping area in the partial image 210 a generated by the image processing unit 104 a. Then, the CPU 101 a controls the projection unit 111 a so as to project the partial image 210 a on which the light reduction processing has been performed. In S313, the projection unit 111 a projects the partial image 210 a, so that the partial image 210 a is displayed on the screen as illustrated in FIG. 1. After the partial image is projected by the projection unit 111 a, the CPU 101 a performs S314. The user can issue a projection end instruction to the projector 100 a by using at least either one of the operation key 107 a and the remote control unit 109 a. The projection end instruction is an instruction to end the processing to display an image on the screen by the multiple projection system.
  • Therefore, in S314, the CPU 101 a determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S314), the CPU 101 a ends the projection processing. When the projection end instruction is not inputted (No in S314), the CPU 101 a returns to S312 and the projection unit 111 a continuously performs projection of the partial image 210 a.
  • After S312 is performed, when the CPU 101 b receives the second coordinate information and the image 200 from the projector 100 a, the CPU 101 b performs S325. In S325, the CPU 101 b controls the image processing unit 104 b so as to generate the partial image 210 b, which is an image to be projected by the projection unit 111 b, based on the image 200 and the second coordinate information. Further, the CPU 101 b controls the image processing unit 104 b so as to perform the light reduction processing on the overlapping area in the partial image 210 b generated by the image processing unit 104 b. Then, the CPU 101 b controls the projection unit 111 b so as to project the partial image 210 b on which the light reduction processing has been performed. In S325, the projection unit 111 b projects the partial image 210 b, so that the partial image 210 b is displayed on the screen as illustrated in FIG. 1. After the partial image is projected by the projection unit 111 b, the CPU 101 b performs S326.
  • While the operation in which the projector 100 b performs S325 has been described, the projectors 100 c and 100 d also perform, in S325, the same operation as that performed by the projector 100 b.
  • The user can issue the projection end instruction to the projector 100 b by using at least either one of the operation key 107 b and the remote control unit 109 b. Therefore, in S326, the CPU 101 b determines whether or not the projection end instruction is inputted. When the projection end instruction is inputted (Yes in S326), the CPU 101 b ends the projection processing. When the projection end instruction is not inputted (No in S326), the CPU 101 b returns to S325 and the projection unit 111 b continuously performs projection of the partial image 210 b.
  • While the operation in which the projector 100 b performs S326 has been described, the projectors 100 c and 100 d also perform, in S326, the same operation as that performed by the projector 100 b.
  • S301 to S314 are processing realized when the CPU 101 a of the projector 100 a, which is the master device, executes a computer program stored in the ROM 102 a. S321 to S326 are processing realized when the CPU 101 b of the projector 100 b, which is the slave device, executes a computer program stored in the ROM 102 b. S321 to S326 are processing realized when the CPU 101 c of the projector 100 c, which is the slave device, executes a computer program stored in the ROM 102 c. S321 to S326 are processing realized when the CPU 101 d of the projector 100 d, which is the slave device, executes a computer program stored in the ROM 102 d.
  • In the first embodiment, it is described that the display patterns that can be selected by using the selection image 400 are the display patterns 401 to 404. However, the display patterns are not limited to these patterns. For example, the display patterns that can be selected by using the selection image 400 may further include display patterns other than the display patterns 401 to 404.
  • In the first embodiment, in the multiple projection system, the master device projects an image for notifying of the display pattern selected by the user and the projection position of the image to be projected by the master device. The slave device projects an image for notifying of the projection position of the image to be projected by the slave device. Thereby, the user can confirm the selected display pattern and the projection positions of the images to be projected by the slave devices that are currently set. For example, when the positions indicated by the images projected by the slave devices do not correspond to the positions illustrated in the display pattern included in the image projected by the master device, the user can determine that there is an error in the input of the IP addresses of the slave devices. In this case, the user can easily determine that the positions associated with the IP addresses of the slave devices do not correspond to the current installation positions of the slave devices, so that the user can correctly modify the input of the IP addresses of the slave devices.
  • Therefore, in the multiple projection system of the first embodiment, it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
  • In the first embodiment, it is described that when S307 is performed by the master device and S322 is performed by the slave devices, in the multiple projection system, images such as the images 700 a to 700 d as illustrated in FIG. 7 are displayed on the screen. However, it is not limited to this. For example, in the first embodiment, when S307 is performed by the master device and S322 is performed by the slave devices, in the multiple projection system, images such as the images 900 a to 900 d as illustrated in FIG. 9 may be displayed on the screen.
  • In this case, in S306, the projector 100 a transmits position information to the projectors 100 b to 100 d and transmits information representing the display pattern selected in S303 to the projectors 100 b to 100 d. Further, in S307, the projector 100 a controls the projection unit 111 a so as to project the image 900 a instead of the image 700 a. In S322, the projector 100 b controls the image processing unit 104 b so as to generate the image 900 b based on the position information and the information representing the display pattern which are received from the projector 100 a. Further, the projector 100 b controls the projection unit 111 b so as to project the image 900 b based on the position information received from the projector 100 a. In the same manner as the projector 100 b, in S322, the projector 100 c controls the projection unit 111 c so as to project the generated image 900 c based on the position information and the information representing the display pattern which are received from the projector 100 a. In the same manner as the projector 100 b, in S322, the projector 100 d controls the projection unit 111 d so as to project the generated image 900 d based on the position information and the information representing the display pattern which are received from the projector 100 a.
  • Second Embodiment
  • Next, a multiple projection system in the second embodiment will be described. In the second embodiment, the description of the same configuration and processing as those in the first embodiment will be omitted.
  • FIG. 10 is a sequence diagram illustrating an example of projection processing performed by the multiple projection system in the second embodiment.
  • S301 to S305, S307 to S314, S321, S325, and S326 in the projection processing in FIG. 10 are the same processing as S301 to S305, S307 to S314, S321, S325, and S326 in the projection processing in FIG. 3, so that the description thereof will be omitted.
  • The projection processing illustrated in FIG. 10 will be described by using an example in which the projector 100 a is set as a master device and the projectors 100 b to 100 d are set as slave devices. Further, the projection processing illustrated in FIG. 10 will be described by using an example in which the display pattern 402 is selected in S303.
  • After performing S305, the CPU 101 a performs S1001. In S1001, in the same manner as in S310 in FIG. 3, the CPU 101 a connects to the projectors 100 b to 100 d through the network IF 105 a based on the IP addresses inputted by the user in S304.
  • After S1001 is performed, the CPU 101 a requests the projectors 100 b to 100 d to transmit resolution information representing the resolution of an image projected by each of the projectors 100 b to 100 d to the projector 100 a. Thereafter, in the same manner as in S323 in FIG. 3, in S1011, the CPU 101 b transmits the resolution information representing the resolution of the image projected by the projector 100 b to the projector 100 a through the network IF 105 b. The projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S1011. The CPU 101 a stores the resolution information received through the network IF 105 a into the RAM 103 a. Thereafter, the CPU 101 a requests the projectors 100 b to 100 d to transmit area information indicating an overlapping area of an image projected by each of the projectors 100 b to 100 d to the projector 100 a. In the same manner as in S324 in FIG. 3, in S1012, the CPU 101 b transmits the area information indicating the overlapping area of the projector 100 b to the projector 100 a through the network IF 105 b according to the request from the projector 100 a. The projectors 100 c and 100 d also perform the same operation as that of the projector 100 b in S1012. The CPU 101 a stores the area information received through the network IF 105 a into the RAM 103 a.
  • When all the resolution information and all the area information are acquired, the CPU 101 a performs S1002. In S1002, the CPU 101 a performs determination processing to determine whether or not the setting performed in S305 is correct by using all the acquired area information.
  • As the determination processing, the CPU 101 a performs first processing, second processing, third processing, and fourth processing.
  • The first processing is processing for determining whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not. The CPU 101 a acquires first information indicating the position of the overlapping area included in the predetermined area 510 a from the RAM 103 a. The first information is stored in the RAM 103 a in advance. The right side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 b, so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located rightward. Further, the lower side of the image projected onto the predetermined area 510 a overlaps with a part of the image projected onto the predetermined area 510 c, so that the first information includes information indicating that the overlapping area included in the predetermined area 510 a is located downward.
  • The CPU 101 a compares the first information with the area information acquired from the projector having the IP address inputted into the input field 522 a, and determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not according to the comparison result. When the first information corresponds with the area information acquired from the projector having the IP address inputted into the input field 522 a, the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct. When the first information does not correspond with the area information acquired from the projector having the IP address inputted into the input field 522 a, the CPU 101 a determines that the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is not correct.
  • The second processing is processing for determining whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not. The CPU 101 a acquires second information indicating the position of the overlapping area included in the predetermined area 510 b from the RAM 103 a. The second information is stored in the RAM 103 a in advance. The left side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 a, so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located leftward. Further, the lower side of the image projected onto the predetermined area 510 b overlaps with a part of the image projected onto the predetermined area 510 d, so that the second information includes information indicating that the overlapping area included in the predetermined area 510 b is located downward.
  • The CPU 101 a compares the second information with the area information acquired from the projector having the IP address inputted into the input field 522 b, and determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not according to the comparison result. When the second information corresponds with the area information acquired from the projector having the IP address inputted into the input field 522 b, the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct. When the second information does not correspond with the area information acquired from the projector having the IP address inputted into the input field 522 b, the CPU 101 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct.
  • The third processing is processing for determining whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not. The CPU 101 a acquires third information indicating the position of the overlapping area included in the predetermined area 510 c from the RAM 103 a. The third information is stored in the RAM 103 a in advance. The upper side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 a, so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located upward. Further, the right side of the image projected onto the predetermined area 510 c overlaps with a part of the image projected onto the predetermined area 510 d, so that the third information includes information indicating that the overlapping area included in the predetermined area 510 c is located rightward.
  • The CPU 101 a compares the third information with the area information acquired from the projector having the IP address inputted into the input field 522 c, and determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not according to the comparison result. When the third information corresponds with the area information acquired from the projector having the IP address inputted into the input field 522 c, the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct. When the third information does not correspond with the area information acquired from the projector having the IP address inputted into the input field 522 c, the CPU 101 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is not correct.
  • The fourth processing is processing for determining whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not. The CPU 101 a acquires fourth information indicating the position of the overlapping area included in the predetermined area 510 d from the RAM 103 a. The fourth information is stored in the RAM 103 a in advance. The upper side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 b, so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located upward. Further, the left side of the image projected onto the predetermined area 510 d overlaps with a part of the image projected onto the predetermined area 510 c, so that the fourth information includes information indicating that the overlapping area included in the predetermined area 510 d is located leftward.
  • The CPU 101 a compares the fourth information with the area information acquired from the projector having the IP address inputted into the input field 522 d, and determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not according to the comparison result. When the fourth information corresponds with the area information acquired from the projector having the IP address inputted into the input field 522 d, the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct. When the fourth information does not correspond with the area information acquired from the projector having the IP address inputted into the input field 522 d, the CPU 101 a determines that the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is not correct.
  • When the determination processing is performed in S1002, the CPU 101 a determines whether the association between the predetermined area 510 a and the IP address inputted into the input field 522 a is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct or not. Further, the CPU 101 a determines whether the association between the predetermined area 510 d and the IP address inputted into the input field 522 d is correct or not. After S1002 is performed, the CPU 101 a performs S1003.
  • In S1003, the CPU 101 a transmits result information indicating the determination result of the determination processing in S1002 and the position information recorded in the RAM 103 a to the projectors 100 b to 100 d through the network IF 105 a.
  • S1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct will be described. In this case, in S1003, the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b. The result information is information indicating that there is an error in the input of the IP address.
  • Further, S1003 performed by the CPU 101 a when it is determined that, for example, the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct will be described. In this case, in S1003, the CPU 101 a transmits the result information and the second position information to the projector having the IP address inputted into the input field 522 b. The result information is information indicating that the input of the IP address is correct. After S1003 is performed, the CPU 101 a performs S307.
  • When the projector 100 b receives the result information and the position information transmitted from the projector 100 a in S1002, the projector 100 b performs S1013. In S1013, the CPU 101 b controls the image processing unit 104 b so as to generate an image based on the position information and the result information received from the projector 100 a.
  • When the result information received from the projector 100 a indicates that there is an error in the input of the IP address, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information and for notifying that there is an error in the input of the IP address. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a. In this case, for example, as illustrated in FIG. 11, the projector 100 b projects an image 1101 onto the screen. The image 1101 includes information 1102 indicating that there is an error in the input of the IP address.
  • When the result information received from the projector 100 a indicates that the input of the IP address is correct, the image processing unit 104 b generates an image for notifying of the position indicated by the potion information. Thereafter, the CPU 101 b controls the projection unit 111 b so as to project the image generated by the image processing unit 104 b to the position indicated by the position information received from the projector 100 a. When the result information received from the projector 100 a indicates that the input of the IP address is correct, the projector 100 b may control the projection unit 111 b so as to project an image indicating that the input of the IP address is correct. In this case, the projection unit 111 b projects an image indicating that the input of the IP address is correct at a position indicated by the position information received from the projector 100 a.
  • While the operation of the projector 100 b has been described, the projectors 100 c and 100 d also perform the same operation as that performed by the projector 100 b. After S1013 is performed, the CPU 101 b performs S325.
  • S1001 and S1002 are processing realized when the CPU 101 a of the projector 100 a, which is the master device, executes a computer program stored in the ROM 102 a. S1011 to S1013 are processing realized when the CPU 101 b of the projector 100 b, which is the slave device, executes a computer program stored in the ROM 102 b. S1011 to S1013 are processing realized when the CPU 101 c of the projector 100 c, which is the slave device, executes a computer program stored in the ROM 102 c. S1011 to S1013 are processing realized when the CPU 101 d of the projector 100 d, which is the slave device, executes a computer program stored in the ROM 102 d.
  • In the multiple projection system in the second embodiment, the master device projects an image for notifying of the display pattern selected by the user and the position of the image to be projected by the master device. Further, the slave device projects an image for notifying of the position of the image to be projected by the slave device.
  • In the second embodiment, the master device determines whether or not there is an error in the input of the IP address of the slave device based on the position of the overlapping area of the image projected by the slave device and notifies the slave device of the determination result. Further, when the slave device is notified by the master device that there is an error in the input of the IP address, the slave device projects an image for notifying information indicating that there is an error in the input of the IP address. Thereby, the user can easily determine that the position associated with the IP address of the slave device does not correspond to the current installation position of the slave device, so that the user can correctly modify the input of the IP address of the slave device.
  • Therefore, in the multiple projection system of the second embodiment, it is possible to cause the user to easily determine whether the association between the IP addresses of the slave devices and the positional relationship of the slave devices is correct or not.
  • It is assumed that the multiple projection system in the second embodiment has the same effect as that of the first embodiment regarding the same configuration and processing as those of the first embodiment.
  • In the second embodiment, the projector 100 a notifies the projectors 100 b to 100 d of the determination result of the determination processing of S1002 and the projectors 100 b to 100 d project the determination result of the determination processing of S1002 which is notified from the master device. However, it is not limited to this.
  • For example, when the projector 100 a receives all the area information, the projector 100 a searches for information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b from among the first information, the second information, the third information, and the fourth information. Thereafter, the projector 100 a changes the setting of the IP address corresponding to the predetermined area 510 b according to the information corresponding to the area information acquired from the projector having the IP address inputted into the input field 522 b. For example, when the IP address of the projector 100 b is inputted into the input field 522 b in S305, the projector 100 a searches for information corresponding to the area information acquired from the projector 100 b from among the first information, the second information, the third information, and the fourth information. When the information corresponding to the area information acquired from the projector 100 b is the second information, the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is correct and then performs S1003. When the information corresponding to the area information acquired from the projector 100 b is the third information, the projector 100 a determines that the association between the predetermined area 510 b and the IP address inputted into the input field 522 b is not correct. Thereafter, the projector 100 a changes the setting of the IP address of the projector 100 b so that the IP address of the projector 100 b is inputted into the input field 522 c. Thereby, the predetermined area 510 c and the IP address of the projector 100 b inputted into the input field 522 c is associated with each other. Thereafter, the projector 100 a determines that the association between the predetermined area 510 c and the IP address inputted into the input field 522 c is correct and performs S1003. The projector 100 a may also perform the same processing for the other predetermined areas 510 a, 510 c, and 510 d.
  • Third Embodiment
  • Next, a multiple projection system in the third embodiment will be described. In the third embodiment, the description of the same configuration and processing as those in the first and the second embodiments will be omitted. In the third embodiment, a multiple projection system will be described where when each of the projectors 100 a to 100 d projects a partial image based on, for example, the display pattern 402, the display pattern is changed to a different display pattern by a user. After the processing of S313 described in FIG. 3 is performed, the CPU 101 a performs change processing illustrated in FIG. 12.
  • When the user operates at least either one of the operation key 107 a and the remote control unit 109 a and thereby the display pattern is changed, the CPU 101 a performs S1201.
  • In S1201, the CPU 101 a receives a change instruction of the display pattern through the user IF 106 a. After S1201 is performed, the CPU 101 a performs S1202. In S1202, the CPU 101 a updates the position information set in S305 based on the display pattern after change indicated by the change instruction of the display pattern and the display pattern before change. Hereinafter, regarding the processing of S1201, a case will be described in which the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 401. Further, a case will be described in which the display pattern before change is the display pattern 402 and the display pattern after change is the display pattern 403.
  • In the display pattern 402, the identification information is “1” and “2” from the left (the first column) of the first row and “3” and “4” from the left (the first column) of the second row. On the other hand, in the display pattern 403, the identification information is “1”, “2”, and “3” from the left (the first column) of the first row and “4”, “5”, and “6” from the left (the first column) of the second row. Therefore, the CPU 101 a updates the position information stored in the RAM 103 a so that the display pattern after change corresponds to the installation positions of the projectors 100 a to 100 d.
  • For example, the CPU 101 a associates the IP address associated with the identification information “3” (the second row, the first column) of the display pattern 402 with the identification information “4” (the first row, the third column) of the display pattern 403 and stores the IP address into the RAM 103 a. Further, the CPU 101 a associates the IP address associated with the identification information “4” (the second row, the second column) of the display pattern 402 with the identification information “5” (the second row, the first column) of the display pattern 403 and stores the IP address into the RAM 103 a. Although no IP address is associated with the identification information “3”, if the IP address for the identification information “6” is set, the CPU 101 a connects to a projector through the network IF 105 a by using the IP address corresponding to the identification information “6”. If the CPU 101 a cannot establish a connection with the projector by using the IP address corresponding to the identification information “6”, it is assumed that no projector to be a connection destination is installed.
  • After S1202 is performed, the CPU 101 a performs S1203. In S1203, the CPU 101 a requests resolution information from the projectors 100 b to 100 d. When the CPU 101 a acquires all the resolution information from the projectors 100 b to 100 d, the CPU 101 a performs S1204. The CPU 101 a determines the resolution of a projector of which resolution information cannot be acquired to be the same as the resolution of another projector. For example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b, the CPU 101 a determines that the resolution of the projector 100 b is the same as the resolution of the projector 100 c adjacent to the projector 100 b. Further, for example, if the CPU 101 a cannot acquire the resolution information from the projector 100 b, the CPU 101 a may determine that the resolution of the projector 100 b is the same as the resolution of the projector 100 a which is the master device.
  • In S1204, the CPU 101 a requests area information from the projectors 100 b to 100 d. When the CPU 101 a acquires all the area information from the projectors 100 b to 100 d, the CPU 101 a performs S1205.
  • In S1205, the CPU 101 a determines whether or not inconsistency occurs between adjacent sides based on the area information acquired in S1204. The adjacent sides are, for example, the right side of the image displayed in the area corresponding to the identification information “1” and the left side of the image displayed in the area corresponding to the identification information “2” in the display pattern 402.
  • When the CPU 101 a determines that inconsistency occurs between adjacent sides (Yes in S1205), the CPU 101 a performs S1206. When the CPU 101 a determines that no inconsistency occurs between adjacent sides (No in S1205), the CPU 101 a performs S1207. In S1206, the CPU 101 a deletes the area information acquired in S324 from the RAM 103 a and stores the area information acquired in S1204 into the RAM 103 a. When the S1206 is performed, the area information is updated.
  • When the display pattern is changed from the display pattern 402 to the display pattern 403 by the user, the number of projectors that project partial images increases. In this case, the CPU 101 a does not perform the processing of S1206 and maintains consistency between the projectors in the multiple projection system by using the area information acquired in S324.
  • After S1206 is performed, the CPU 101 a performs S1207. The processes of S1207 to S1209 are respectively the same as the processes of S311 to S313 illustrated in FIG. 3. FIG. 13A is a diagram illustrating an image 1301 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 401. FIG. 13B is a diagram illustrating an image 1302 displayed on the screen when the display pattern is changed from the display pattern 402 to the display pattern 403.
  • In the multiple projection system in the third embodiment, when the display pattern is changed by the user, the master device resets the projection positions of the slave devices based on the display pattern before change and the display pattern after change.
  • Thereby, in the multiple projection system in the third embodiment, it is possible to automatically set the projection positions of the slave devices by changing the display pattern by the user.
  • It is assumed that the multiple projection system in the third embodiment has the same effect as that of the first and the second embodiments regarding the same configuration and processing as those of the first and the second embodiments.
  • In the description of the first, the second, and the third embodiments, the master device is the projector 100 a and the slave devices are the projectors 100 b to 100 d. However, it is not limited to this. For example, the master device may be the projector 100 d and the slave devices may be the projectors 100 a to 100 c.
  • OTHER EMBODIMENTS
  • Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
  • While the present disclosure has been described with reference to exemplary embodiments, it is to be understood that the disclosure is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.

Claims (46)

What is claimed is:
1. A method for displaying a combined image by displaying a plurality of images by a plurality of display apparatuses,
wherein one of the plurality of display apparatuses communicates with another of the plurality of display apparatuses,
wherein the one display apparatus displays an image for indicating a positional relationship between the plurality of images to be displayed by the plurality of display apparatuses, and
wherein the another display apparatus displays information for notifying an area of an image to be displayed by the another display apparatus.
2. The method according to claim 1, wherein
the another display apparatus displays, by the communication between the one display apparatus and the another display apparatus, information for notifying a position of the area of the image to be displayed by the another display apparatus in the combined image.
3. The method according to claim 1, wherein
the one display apparatus displays a setting image for setting a plurality of areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
4. The method according to claim 3, wherein
the one display apparatus displays an image related to information for identifying positions of the plurality of areas.
5. The method according to claim 3, wherein
the one display apparatus displays a plurality of area images corresponding to the plurality of areas.
6. The method according to claim 5, wherein
the one display apparatus displays information for identifying the plurality of areas, on the plurality of area images.
7. The method according to claim 1, wherein
the one display apparatus displays an image including identification information for the plurality of display apparatuses and information indicating areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
8. The method according to claim 7, wherein
the identification information is an IP address.
9. The method according to claim 7, wherein
the one display apparatus communicates with the another display apparatus based on setting done by using the image.
10. The method according to claim 1, wherein
the one display apparatus communicates with the another display apparatus through a wireless LAN.
11. The method according to claim 1, wherein
the plurality of display apparatuses is a plurality of projectors.
12. A display apparatus used for displaying a combined image by combining a plurality of images displayed by a plurality of display apparatuses comprising:
at least one CPU; and
at least one memory storing a program which, when executed by the at least one CPU, causes the display apparatus to:
display an image for indicating a positional relationship between the plurality of areas;
communicate with another of the plurality of display apparatuses;
control the communication with the another display apparatus so that the another display apparatus displays information for notifying an area of an image to be displayed by the another display apparatus in the combined image.
13. The display apparatus according to claim 12, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to control the communication with the another display apparatus so that the another display apparatus displays information for notifying a position of the area of the image to be displayed by the another display apparatus in the combined image.
14. The display apparatus according to claim 12, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display a setting image for setting a plurality of areas of the plurality of images to be displayed by the plurality of display apparatuses.
15. The display apparatus according to claim 14, wherein
the setting image includes an image related to information for identifying positions of the plurality of areas.
16. The display apparatus according to claim 14, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display a plurality of area images corresponding to the plurality of areas.
17. The display apparatus according to claim 16, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display information for identifying the plurality of areas on the plurality of area images.
18. The display apparatus according to claim 12, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display an image including identification information for the plurality of display apparatuses and information indicating areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
19. The display apparatus according to claim 18, wherein
the identification information is inputted by user operation.
20. The display apparatus according to claim 18, wherein
the identification information is an IP address.
21. The display apparatus according to claim 12, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to perform control to display information about an area of an image to be displayed by the display apparatus in the combined image.
22. The display apparatus according to claim 12, wherein
the plurality of display apparatuses is a plurality of projectors,
the program, when executed by the at least one CPU, further causes the display apparatus to project an image.
23. A method for displaying a combined image by displaying a plurality of images by a plurality of display apparatuses,
wherein one of the plurality of display apparatuses communicates with another of the plurality of display apparatuses, and
wherein the another display apparatus displays, by the communication between the one display apparatus and the another display apparatus, information for notifying an area of an image to be displayed by the another display apparatus in the combined image.
24. The method according to claim 23, wherein
the another display apparatus displays, by the communication between the one display apparatus and the another display apparatus, information for notifying a position of the area of the image to be displayed by the another display apparatus in the combined image.
25. The method according to claim 23, wherein
the one display apparatus displays a setting image for setting a plurality of areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
26. The method according to claim 25, wherein
the setting image includes an image related to information for identifying positions of the plurality of areas.
27. The method according to claim 23, wherein
the one display apparatus displays a pattern image for indicating a positional relationship between the plurality of areas.
28. The method according to claim 27, wherein
the one display apparatus displays information for identifying the plurality of areas on positions corresponding to the plurality of areas in the pattern image.
29. The method according to claim 23, wherein
the one display apparatus displays an image including identification information for the plurality of display apparatuses and information indicating areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
30. The method according to claim 29, wherein
the identification information is inputted by user operation.
31. The method according to claim 29, wherein
the identification information is an IP address.
32. The method according to claim 25, wherein
the communication is performed based on setting done by using the setting image.
33. The method according to claim 23, wherein
the communication is performed through a wireless LAN.
34. The method according to claim 23, wherein
the plurality of display apparatuses is a plurality of projectors.
35. A display apparatus used for displaying a combined image by combining a plurality of images displayed by a plurality of display apparatuses comprising:
at least one CPU; and
at least one memory storing a program which, when executed by the at least one CPU, causes the display apparatus to:
communicate with another of the plurality of display apparatuses;
control the communication with the another display apparatus so that the another display apparatus displays information for notifying an area of an image to be displayed by the another display apparatus in the combined image.
36. The display apparatus according to claim 35, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to control the communication with the another display apparatus so that the another display apparatus displays information for notifying a position of the area of the image to be displayed by the another display apparatus in the combined image.
37. The display apparatus according to claim 35, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display a setting image for setting a plurality of areas of the plurality of images to be displayed by the plurality of display apparatuses.
38. The display apparatus according to claim 37, wherein
the setting image includes an image related to information for identifying positions of the plurality of areas.
39. The display apparatus according to claim 35, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display a pattern image for indicating a positional relationship between the areas corresponding to the plurality of display apparatuses.
40. The display apparatus according to claim 39, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display information for identifying the plurality of areas on positions corresponding to the plurality of areas in the pattern image.
41. The display apparatus according to claim 35, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to display an image including identification information for the plurality of display apparatuses and information indicating areas of the plurality of images to be displayed by the plurality of display apparatuses in the combined image.
42. The display apparatus according to claim 41, wherein
the identification information is inputted by user operation.
43. The display apparatus according to claim 41, wherein
the identification information is an IP address.
44. The display apparatus according to claim 37, wherein
the communication is performed based on setting done by using the setting image.
45. The display apparatus according to claim 35, wherein
the program, when executed by the at least one CPU, further causes the display apparatus to perform control to display information about an area of an image to be displayed by the display apparatus in the combined image.
46. The display apparatus according to claim 35, wherein
the plurality of display apparatuses is a plurality of projectors,
the program, when executed by the at least one CPU, further causes the display apparatus to project an image.
US16/377,914 2015-04-02 2019-04-08 Projection device and control method therefor Abandoned US20190235823A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/377,914 US20190235823A1 (en) 2015-04-02 2019-04-08 Projection device and control method therefor

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
JP2015-076125 2015-04-02
JP2015076125A JP6275075B2 (en) 2015-04-02 2015-04-02 Projection apparatus, control method, and program
US15/084,325 US9875074B2 (en) 2015-04-02 2016-03-29 Projection device and control method therefor
US15/844,352 US10394514B2 (en) 2015-04-02 2017-12-15 Projection device and control method therefor
US16/377,914 US20190235823A1 (en) 2015-04-02 2019-04-08 Projection device and control method therefor

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/844,352 Division US10394514B2 (en) 2015-04-02 2017-12-15 Projection device and control method therefor

Publications (1)

Publication Number Publication Date
US20190235823A1 true US20190235823A1 (en) 2019-08-01

Family

ID=57015198

Family Applications (3)

Application Number Title Priority Date Filing Date
US15/084,325 Active US9875074B2 (en) 2015-04-02 2016-03-29 Projection device and control method therefor
US15/844,352 Active US10394514B2 (en) 2015-04-02 2017-12-15 Projection device and control method therefor
US16/377,914 Abandoned US20190235823A1 (en) 2015-04-02 2019-04-08 Projection device and control method therefor

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/084,325 Active US9875074B2 (en) 2015-04-02 2016-03-29 Projection device and control method therefor
US15/844,352 Active US10394514B2 (en) 2015-04-02 2017-12-15 Projection device and control method therefor

Country Status (2)

Country Link
US (3) US9875074B2 (en)
JP (1) JP6275075B2 (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6798108B2 (en) * 2016-01-20 2020-12-09 セイコーエプソン株式会社 Image projection system and control method of image projection system
JP6707870B2 (en) * 2016-01-20 2020-06-10 セイコーエプソン株式会社 Projection system and projection position detection method
JP6946647B2 (en) * 2017-01-11 2021-10-06 セイコーエプソン株式会社 Display device and control method
JP7129149B2 (en) * 2017-06-29 2022-09-01 キヤノン株式会社 Projector and its control method
US11093197B2 (en) * 2017-07-31 2021-08-17 Stmicroelectronics, Inc. System and method to increase display area utilizing a plurality of discrete displays
CN107959836A (en) * 2017-11-15 2018-04-24 苏州佳世达光电有限公司 A kind of projecting method and optical projection system
JP6770502B2 (en) 2017-11-22 2020-10-14 キヤノン株式会社 Communication devices, display devices, their control methods, programs and display systems
DE102018112873A1 (en) * 2018-05-29 2019-12-05 Ebm-Papst Mulfingen Gmbh & Co. Kg Auto addressing with position determination of bus users
CN109062531B (en) * 2018-08-15 2021-08-27 京东方科技集团股份有限公司 Spliced screen, display method of spliced screen and display control device
JP7238298B2 (en) * 2018-09-03 2023-03-14 セイコーエプソン株式会社 Projectors and projector control methods
CN112328193A (en) * 2020-09-09 2021-02-05 深圳Tcl新技术有限公司 Multi-terminal screen combination method, device, equipment and computer storage medium
CN113590067A (en) * 2021-07-28 2021-11-02 深圳创维-Rgb电子有限公司 Screen projection control method, system, device and computer readable storage medium

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2501965B2 (en) * 1991-02-12 1996-05-29 パイオニア株式会社 Data transfer control device
KR100465173B1 (en) * 2002-11-18 2005-01-13 삼성전자주식회사 Image displayer with separating screen function
JP2006284990A (en) * 2005-04-01 2006-10-19 Matsushita Electric Ind Co Ltd Image transmission multi-display system
KR101234156B1 (en) * 2006-06-14 2013-02-15 삼성전자주식회사 Method for providing external-input list using item-grouping and video apparatus thereof
JP2006338044A (en) * 2006-07-24 2006-12-14 Sony Corp Multi-display device, and multi-display control method, and computer program
US7901094B2 (en) * 2007-01-24 2011-03-08 Seiko Epson Corporation View projection: one-touch setup of light displays on arbitrary surfaces
JP5515351B2 (en) * 2009-03-24 2014-06-11 セイコーエプソン株式会社 Image output apparatus, control method, and control program
JP2010262201A (en) * 2009-05-11 2010-11-18 Seiko Epson Corp Projector system, projector, method and program for controlling projector
KR101694159B1 (en) * 2010-04-21 2017-01-09 엘지전자 주식회사 Augmented Remote Controller and Method of Operating the Same
JP5641820B2 (en) * 2010-08-25 2014-12-17 キヤノン株式会社 Projection type display system and projection type display device
JP2014006357A (en) * 2012-06-22 2014-01-16 Seiko Epson Corp Projector, image display system, and control method of the projector
JP2014107713A (en) * 2012-11-28 2014-06-09 Seiko Epson Corp Operation method, operation program and operation apparatus
US9041863B2 (en) * 2013-08-30 2015-05-26 Xiaomi Inc. Electronic device and method for displaying resources

Also Published As

Publication number Publication date
US20160291919A1 (en) 2016-10-06
US9875074B2 (en) 2018-01-23
JP6275075B2 (en) 2018-02-07
US20180107443A1 (en) 2018-04-19
JP2016197146A (en) 2016-11-24
US10394514B2 (en) 2019-08-27

Similar Documents

Publication Publication Date Title
US10394514B2 (en) Projection device and control method therefor
US20170142382A1 (en) Communication apparatus, method of controlling communication apparatus, non-transitory computer-readable storage medium
JP2014107713A (en) Operation method, operation program and operation apparatus
JP5870586B2 (en) Projector control device, display device, and program.
US10754237B2 (en) Control apparatus, control method, projection system, and storage medium
US20160027402A1 (en) Wireless communications system, and display apparatus
US10303419B2 (en) Information processing system, display processing apparatus, display processing method, and recording medium
US10768884B2 (en) Communication apparatus, display apparatus, control method thereof, storage medium, and display system for configuring multi-display settings
US20200169706A1 (en) Projection control apparatus, control method of the same, and projection system
US20140211168A1 (en) Image projection apparatus, control method, recording medium, and projection system
JP2015169940A (en) Multi-projection system, information processor and program
US10009583B2 (en) Projection system, projection apparatus, information processing method, and storage medium
US10484550B2 (en) Display control apparatus that displays preview image, control method therefor, and storage medium
JP2017135490A (en) Display system, display device, control device, and method for controlling these
JP2020178221A (en) Projection control device, projection control method, and program
JP2006270461A (en) Image formation apparatus and method
JP6825480B2 (en) Information information system, display processing device, display processing method, and display program
US9881363B2 (en) Image processing device, projector and image processing method for correcting a displayed image
US11206379B2 (en) Controlled apparatus and control method thereof
JP7129149B2 (en) Projector and its control method
JP2017073717A (en) Information processing unit, information processing method, computer program and recording medium
US20190327457A1 (en) Projection control apparatus and projection control method
JP6883196B2 (en) Processing equipment, information processing equipment, processing systems, processing programs and information processing programs
CN115343898A (en) Projection system and projection image superposition method
JP2012203311A (en) Information processing system, information processing apparatus and program

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION