This paper presents a tool to transform a smartphone into a smart eyewear, named “CARTON”, following a Do-It-Yourself (DIY) approach. The hardware prototype is made with very simple materials and regular tools we could find anywhere. It also includes a Software Development Kit (SDK) with samples in order to easily adapt or develop new mobile app compatible with this kind of device. By providing everything open-source and open-hardware, we intend to solve the reachability of technologies related to smart eyewear and aim to accelerate research around it. Users experiments were conducted in which participants were asked to create, by themselves, the CARTON’s hardware part and perform usability tests with their own creation. Qualitative user feedback and quantitative results prove that CARTON is functional and feasible by anyone, without specific skills.
Nowadays, smart eyewear and others assimilate Optical Head- Mounted Display (OHMD) as Google Glass, Vuzix, Ora-X, among others, are still very expensive or worst, sometimes even unavailable or hard to obtain depending on the location (e.g., limited to one or few countries). This lack of availability slows the broad use of such technology by the population. These issues make more difficult to conduct research and develop prototypes with smart eyewear. For instance, conducting research on collaborative augmented reality systems for schools, in an entire class would be currently hard and costly to obtain a numerous amount of smart eyewear devices.
Very few studies focused on solving this problem. Some projects intended to reduce costs by applying smartphones to create an inexpensive AR (Augmented Reality) headset. The use of a smartphone to create an accessible smart eyewear is considered a key strategy for two reasons:
It provides everything needed for the “smart” part of this eyewear (e.g., chip, screen, camera, battery, connectivity and is usually packed with plenty of sensors).
It is widely available. Just for Android in September 2015, Google claims there were 1.4 billion 30-day active devices around the world .
Unfortunately, previous research applying the smartphone approach for developing smart eyewear has not solved the problem yet. Whether their solutions were not similar enough to smart eyewear, but closer to a virtual reality headset , or they lacked reachability [6, 15]: considering the process of creation is not open or it includes tools/materials not yet widely available.
In this paper, by smart eyewear, we mean any device which can carry out many functions of a mobile computer, but has to be worn on the head, implying the use of eyes. Therefore, a smart eyewear is not exclusively related to augmented reality technology, and could be applied for other experiences.
In this study, we aim to solve these issues by designing a DIY smart eyewear prototype (shown in Figure 1) which requires only simple material and regular tools that mostly everyone has access around the world, at least where smartphones are present. By using standard material and tools we aim to create favorable conditions to knowledge-sharing with instructions easy to follow . Information and knowledge exchange is a core value to the motivations for contributing in the DIY community : allowing creators and contributors to receive feedback, educate others and showcase projects, among others. Highlighting its value, DIY practices are also popular among CHI researchers . Therefore, in order to make it truly reachable, we naturally decided to make it open-hardware. Thus, all guides and blueprints are available directly on the Internet (https://mobicarton.github.io).
“CARTON”, as a complete toolkit also includes an open-source SDK. The SDK aims to accelerate and simplify the adaptation, or development, of existing and new mobile app compatible with our device (e.g., mostly any application dedicated to Google Glass and assimilate could be adapted for CARTON thanks to similar features and characteristics). The open-source community based on knowledge-sharing are quite similar to the DIY community sharing some identical cores. Furthermore, open-source practices have already proved to be effective in software development . Almost as old as the early age of software development, the open- source communities have grown and are very widespread around the world, creating a huge number of tools and new technology in order to improve their effectiveness.
In our solution, to complete the open-source SDK, the toolkit also contains a full documentation and two sample apps used during our experiments. Indeed, we conducted experiments with 16 participants to confirm that this DIY smart eyewear is feasible, usable and functional following the guide, the blueprint and its utilization with real applications. Therefore, the contribution of this project is double: (i) providing blueprints and guidelines to build a low-cost, open source and easy to do DIY smart eyewear devices and (ii) providing an SDK to support the development and the adaptation of mobile apps for eye-wearable. At last, CARTON project allows to end users an easy access to another kind of multimedia (e.g., interface/interaction related to smart eyewear) just with a mobile phone.
This paper is structured as follows. The first section presents a review of the related work around the development of smart eyewear. The second section presents the CARTON project with a description of its hardware and the developed software, including a description of the sample applications that goes with our SDK. The third section presents the evaluation of our solution, followed by the presentation of the results. Finally, we conclude this paper with future work and conclusion.
2. RELATED WORK
Motivated by the inaccessibility of augmented and virtual reality technologies for developing countries’ population, Amer et al.  were ones of the first researchers to present affordable altered perspectives to make augmented (and virtual) reality technology. They presented a tool similar to the well-known Google Cardboard (DIY virtual reality headset)  but with foam board, welding goggles and concave lenses. With the help of stereoscopy technique and smartphone’s camera, their device can create a virtual or augmented reality experience, among others. The major problem of this strategy is that the entire view of the user is limited to the focal of the smartphone’s camera. Also, the reality is altered by the smartphone’s screen and camera characteristics (density, color restitution), feeling less natural. Another limitation of this tool is that their system requires a large amount of initial manual adjustment before use.
After this first initiative, De Angeli and O’Neill  outlined their study in developing an AR headset with low purchase and maintenance costs. Their work is similar to a product previously presented by the company Seebright . In their study, a lightweight frame that holds a smartphone above the user’s eye line was applied. A piece of Plexiglass was added in front of the user’s eye line which refracts the light of the smartphone’s screen, therefore making it possible to add a virtual layer to the reality. Since the rear camera’s phone faces the ceiling, a piece of mirror was added on the top of the headset, aligning the camera’s sight with the eye line. The authors conducted a study to test the visual capacity of the tool, depending on light condition of the environment and smartphone’s screen brightness. The results showed that their headset, using smartphones, delivers very good results with ambient illumination similar to interior environment. Unfortunately, this tool has a huge limitation in relation to its reachability to everyone. Despite the fact that the material does not have an excessive cost, in their last version they used tools which are currently not accessible everywhere, like a 3D printer. Although adaptations on mobile apps are mandatory to use their solution, no SDK is provided. But overall, and most important, all of these previous works did not provide their creation openly to the community, which is unfavorable to its reachability.
Based on an analysis of the related works we propose the CARTON project, which aims a broader use of smart eyewear following a DIY approach.
Aiming to ensure our eyewear’s reachability, we focused our design in employing materials that were easy to obtain and affordable. CARTON is made of very simple materials and applying regular tools. The complete list of needed tools and material is shown in Table 1.
3.1.1 Design Strategies
Inspired in more consolidated products, such as Meta Vision 2 , and a previous research on an inexpensive augmented reality headset , we use Pepper’s ghost effect  to add a layer on the reality. Using this effect, the smartphone’s screen is projected directly into the transparent plastic sheet (as seen in Figure 2), which is positioned in front of the user’s eyes. Since the plastic sheet is perfectly transparent (such as a glass lenses), the real world look through it remains unchanged and normal. An important factor to be taken into consideration in this approach is that the transparent plastic sheet should not be thicker than 0.5mm (0.020 in.), otherwise the projection will appear twice, one on each side, and the quality will be affected (with a blurred effect). Another solution to avoid this issue is to apply an anti-reflective transparent material, but unfortunately, this material is not easily available, hence against the philosophy of this project. This step is very sensible because if the transparent plastic sheet is too thin, then it can bend and the result is distorted. As a strategy to solve this issue, we added to the design a flat stick in order to prevent the transparent plastic sheet bend. This transparent plastic sheet can be found in any regular stationery shop.
In this current configuration, as shown in Figure 3, the rear camera of the mobile phone faces the ceiling. We use a small mirror piece to track the view in front of the user’s eyes, it is placed on the top of the mobile phone at 45°. We designed the prototype so that the mirror is separated from the main part in another module for three reasons:
Tracking the user’s view is optional for many use cases. For instance, to receive context-awareness notification we could use other mobile’s features and sensors (e.g., GPS, Wi-Fi, accelerometer...).
The mirror is probably the hardest piece of material to acquire, so we can still create and use the CARTON without it.
Due to multiple rear camera’s position (depending on the model of the mobile, the position is usually top-left, but sometime top center), thus the separate module makes the CARTON device adaptable to those differences.
Taking into account facial anthropometric , we tilted down the main structure to 15°, because the forehead is not perfectly perpendicular to the straight-ahead look of the eyes. Adding to it, this strategy also contributed to correcting the upper position of the smartphone and its camera from the horizontal eye line, as shown in Figure 3.
Besides the materials previously discussed, extra articles were added to the toolkit aiming to improve its design. The sponge acts as a cushion cover to make this wearable tool more comfortable. The rubber band has been placed to avoid the smartphone to fall out from the carton prototype. The utility stretch strap is there to fix it on the head, adding the hand-free ability to this project. In overall, the total cost of the material, to produce one device, does not exceed $9 USD. Adding to it (only if needed) the additional cost of the tools shown in Table 1, the maximal cost is around $13 USD, depending on the shop and location. However, this kind of tools and materials are usually already owned or can be found for free, just as the main material: cardboard.
Almost all the blueprint has been designed using the open-source software LibreOffice Draw or Inkscape and it has been initially inspired by the Google Cardboard v1. After several design iterations, it evolved to the blueprint presented in Figure 4. During the designing of the blueprint, to accelerate each iteration process, we have been to a fabrication laboratory (FabLab) to laser cut the cardboard. For that reason, even if it is not the main purpose of this project, we also released a version of the blueprint adapted to laser-cut. The final blueprint is on scale and adapted to be printed on regular paper size Letter or A4 format.
Our motivation to create an SDK with two samples mobile apps is deeply linked with its reachability. Instead of just creating a unique demonstration app as a proof of concept we considered necessary to provide an easy way to develop new experience using CARTON devices. Indeed, some coding adaptations to the apps are necessary to be compatible (e.g., to respect some size and margin), therefore it would be considered as a boilerplate code without the SDK. Everything has been created natively with Android Studio 1.5 and Android SDK version 23.
The SDK has been created in order to facilitate the development, or adaptation of mobile app compatible with the CARTON device. It is available for the platform Android 4.1+ and can be downloaded from the project related GitHub repository (https://github.com/mobicarton). In addition to examples, the SDK is also accompanied with documentation and tutorials, which helps his integration. The CARTON SDK includes the following features:
Auto-adaptive screen. It configures the smartphone’s screen automatically to the correct size (width, height), position (margin—left and top) and brightness. The current position size has been chosen considering a wide difference of mobile phone screens’ size between 3.5 and 6 inches. It limits the usable screen zone to 30x60 mm with a 10 mm margin from top and left as shown in Figure 5. We measure one of the smallest popular smartphones (iPhone 4 with 3.5 inches’ screen) and choose those limits based on the minimal. We made sure it fit inside. A design choice made to avoid different experience depending on the user’s phone. Then the brightness is set to maximum to make it more visible . Finally, a mirrored effect is added to the whole interface because of the Pepper’s ghost effect, otherwise everything is horizontally reversed.
A default launcher. This provides a standard launcher to get information on how to place the phone inside the CARTON device. Furthermore, it includes a button that allows users to launch the app without CARTON, the mirrored feature is automatically deactivated, a functionality that is particularly convenient for developers. It also provides a link directly to the construction guide for people without CARTON who wants to build their own.
Head Gesture Recognition API. It helps to integrate a spherical tracking and a head gesture recognition such as tilt and nod. These two gestures are considered as a natural and intuitive interaction, which can be performed without disturbing a primary activity . We use high-level sensor fusion from Android that mixes different sensors: accelerometer, gyroscope and magnetometer. To recognize tilting (right or left) or nodding (up and down), the API measures how far user inclines her/his head with a threshold (15°) and how long does it take to go back to origin (1000 milliseconds) as shown in Figure6 (left). These parameters were chosen considering internal tests, and have been refined after the pilot test.
Finger Gesture. The smartphone’s screen is still reachable to the user to conduct touch gestures. Therefore, we included a layer over the interface to interact with common patterns: swipe/fling left, right and forward/backward (top/down) as shown in Figure 6 (right). Quick gestures have been chosen to avoid causing a lack of interactivity, with the finger cutting off the user’s view. The Finger Gesture feature was added to provide a better accessibility of the device (e.g., people with disability who couldn’t use Head Gesture or Voice Recognition).
Voice Recognition. The native Android speech recognition service is used to allow the user to easily interact with the voice.
3.2.2 Sample (Demos) Two mobile apps have been created using the CARTON SDK. They are also openly available on the GitHub repository. The first application, which is the main one, called “CARTON”, include six features accessible from a menu composed of simple tiles. This app is available for Android 4.2.x, or more, because of a used component added in API 17 (TextClock). We can navigate through it with different kinds of interactions: fingers, head gesture and voice commands. Directly from the main menu (as shown in figure 7) you can access:
Clock. A simple tile providing time.
Compass. The compass app of Google Glass adapted for CARTON with our SDK. The aim of this adaptation is to show that we can easily create and experience something similar to a consolidated smart glass using our tool.
Live Subtitle. Use the Voice Recognition feature from CARTON SDK to create subtitles live-on. Its aim is to help hearing-impaired people, by listening to conversation then transcript it on the air as subtitles.
Origami. Five origami diagrams are included (Frog, Mouse, Tulip, Lily and Cranevar), all of them under a Creative-Common license. They have been adapted to provide guides of each step in the user’s view and uses hand-free control to navigate through the different steps.
Tutorial. There is a tutorial automatically started when the app launch for the first time, but we can access it again directly from the menu. Here we can learn how to interact with CARTON, using some basic finger patterns, head gesture recognition and voice recognition.
Compatible app. It retrieves the name and the description of any app installed on the current phone that is compatible with CARTON. It allows a direct launch from this main app, for example we can launch the second sample that way without leaving this one.
The second application, named “Poster Target”, is used to create an AR (Augmented Reality) experience with the camera’s phone. The view of the phone camera is projected to the screen of CARTON using the top mirror (as seen in Figure 8), then the user is able to focus at several selected printed posters and get additional information about them. The poster recognition is supported by the AR technology Vuforia v5 . We use the Image Target feature in order to provide tracking and recognition of the posters. When a poster is recognized and tracked the app shows to the user, through the CARTON screen, some textual information, like a simple description of the poster in a virtual text box. The posters used for this sample are a courtesy of NASA/JPL-Caltech, dealing with space exploration and called the Visions of the Future (http://jpl.nasa.gov/visions-of-the-future/). This app is built in order to check two characteristics of CARTON:
Hardware: use of the camera’s phone and so the mirror module.
Software: CARTON SDK and its compatibility with Augmented Reality technologies (e.g., Vuforia).
To make sure the sample apps and the CARTON prototype work properly with multiple devices, we successfully conduct internal tests with popular android phones, as shown in Table 2. iPhone hasn’t been tested yet because the SDK is currently available only for Android OS.
The main objective of the evaluation is to answer two main questions to make sure this project is appropriate and reachable to a wide community: Is it feasible? Is it functional? Firstly, we aimed to evaluate the whole Do-It-Yourself creation process, including the quality of blueprints, guide, tools and materials used. Furthermore, in a second part, we aimed to evaluate CARTON’s functionality and usability with sample apps.
We recruited 16 participants for the experiments, after publishing a recruitment poster in a university and in a video game co- working space. Therefore, the user groups were composed mostly of students (7) or people related to the video game industry (4), 12 men and 4 women aged between 25 and 56. In exchange for their participation in the study, participants received their CARTON creation and $10, as a lump sum refund for public transport. Since they also kept their own creation, we hoped it made participants more conscientious with the assigned tasks. Our aim was attracting people without specific skills which could advantage them to create a CARTON prototype. Only 6 of the participants already created some DIY project before. Few confess, in the survey accompanying the evaluation, that they were not particularly handymen. 3 participants had knowledge and skills to develop mobile app, so they were interested in using it after the experiments, but also two of them did not own a smartphone.
During one week we managed 4 sessions, with 4 participants each. In each session we provided all the material and tools required that we found easily in regular shops. Out of the 16 participants, 4 took part in a pre-test (the first session) in order to update the guide and blueprint. Thus, the main evaluation group included 12 participants, which is considered enough participants to identify around 95.0% of significant usability problems .
Each session lasted around two and a half hours and was organized in two parts. In the first part, participants were asked to follow the guidelines and blueprint to create their own CARTON. In the second part, participants tested their creation using the sample demos.
4.1.3 Construction Process
During this first part of the experimentation, participants created their own CARTON. The aim of this activity was to reproduce a situation as if they were at home, following guidelines and blueprints that they could find on the Internet. Participants were encouraged making comments (i.e. Think aloud method) during the experimentation. When they finished the creation of their CARTON device, they had to answer the first part of a questionnaire that includes 22 questions (mainly about the guide, the blueprint, the tools, their feelings). To make the experimentation process more similar to what we expect to happen in real life, we intentionally choose a regular to low quality of cardboard, since it is the most available in our daily life.
4.1.4 Usability Tests
During the second part of the experimentation, participants had to use their own creation with a mobile phone we provided to them: Nexus 6 or Nexus 6P, both running on Android 6.0.1. We didn’t let them use their own smartphone in order to reproduce the same experience across participants with similar characteristics. For this activity, we used the two sample apps compatible with CARTON that were presented in the previous section. Participants had to go through the tutorial in the main app and navigate in the menu with fingers or head gesture. Then they were asked to launch the second app to test the mirror module. Finally, participants answered the last part of the questionnaire, with 29 questions about their experience (about the comfort, the display, few other characteristics and their feeling).
The results from the participants’ observation and questionnaire show positive outcomes regarding the CARTON construction process. Most participants could successfully create their own smart eyewear using the CARTON toolkit and all of them have been proud of using a product they build by themselves. Only one participant could not finish his construction properly, as its mirror module could not fit well. Except for this participant, all other participants affirmed to get some pleasure to use their smart eyewear.
Furthermore, participants also rated positively their experience in the creation process and the provided user guide, blueprint and tools. Participants’ satisfaction in the creation process was rated as 4.33/5 (1 unsatisfied/5 satisfied). Regarding the user guide and blueprint, most participants considered both as good, 91.7% (11/12) and 83.3% (10/12) respectively. The recommended tools’ choice (that corresponds to: utility knives, scissors, ruler and glue stick) was also rated positively, 100% of the participants considered it suitable/adequate. Besides showing the value of the CARTON to allow users, without specialized skill, to build a low-cost, open source and DIY smart eyewear, the results of the evaluation also allowed us to identify improvement aspects in relation to difficulties to build the device and its robustness. Participants affirmed that building the headset was not so easy 2.66/5 (1: hard, 5: easy). However, beyond the questionnaire, we notice during the process of construction that most participants tend to not read the guide and rely more on the blueprint. This could sometimes be an issue with forgotten and broken parts that made the construction more difficult, lengthen the process and also weakened the CARTON. Possible strategies to facilitate the construction process could be providing a video tutorial, and performing some improvements to the actual guide and blueprint.
In relation to the device robustness, even if 66.7% (8/12) of the participants think their CARTON is not robust mainly because of the cardboard, 83.3% (10/12) think the material’s choice is suitable. Both of the two “inadequate” answers were due to the quality of the cardboard, participants said the cardboard was “coarse for a fine cut” and “too thick”. We intentionally selected a low quality cardboard for reachability reasons previously explained, therefore this problem could be easily minimized. As a reflection of their judgment on robustness, 66.7% (8/12) of the participants affirmed they would prefer to create their own version with a 3D printer. Their motivations for that were: a need for “more robust” (said by 5 participants) but also “more accurate” (3 participants) and two of them for the “aesthetic”. We consider that a 3D printing solution is not as highly customizable by a wide community, as the cardboard solution of CARTON, because it requires more specialized knowledge and skills (e.g., in 3D modeling) to update the prototype. With our cardboard solution we ensure flexibility with customization, adaptation and personalization to satisfy more users  without adding extra cost.
The results of the experimentation highlighted few improving aspects regarding the CARTON’s comfort and ergonomics. Measure of the easiness to put on and take off the device was rated 3.33/5 (1: hard, 5: easy), which is correct (close to easy). Sometimes due to the chosen material, the utility stretch straps were judged “too small”, not big enough for some heads, or it could stick with long hairs. The headset comfort was judged to 3.5/5 (1: bad, 5: good), still correct (close to good), but not a perfect score that is linked with the precedent minor and correctable default where the small utility stretch was judged “too tight”, or the sponge “too harsh”.
Half of the participants felt the weight of the device, but only one pointed “cumbersome” as a comfort issue. This issue could be easily minimized with another lighter smartphone than those used during the experiment. Nexus 6P and Nexus 6 are both of the heaviest phone with 178 g (6.28 oz.) and 184 g (6.49 oz.). The headset weight 85 g (3.0 oz.), so the total was 263 g and 269 g. To compare with Google Glass which weighs only 36 g (1.27 oz.), it is more than 7 times more. If we change for a lighter phone like Nexus 5X which weight 136g (4.80 oz.), we would gain around 20%.
The size of the display was judged large enough by 66.7% (8/12) of the participants and its render quality get 3.25/5 (1: bad, 5: good), which is correct (above average and closer to good) particularly considering that 2 participants had some trouble because their eyes have refractive errors, therefore they scored low this quality. Also, due to a difficulty in the construction process, two other participants had issues with a distorted plastic sheet that makes the result a bit blurred. As a consequence of the stretch being too tight, half of the participants felt physical discomfort and 75% of them felt a bit of fatigue but still tolerable after using the headset. However, these outcomes concur with the results of a previous research on smart eyewear (commercialized HMD) . Indeed, their results showed that even that there were no significant quantitative results proving a visual fatigue of the users (such as blink rate, etc.) participants answered subjectively they felt such a feeling. Also we point out that the answer of our survey might be influenced because the usability tests took place at the end of the experiments, straight after almost two hours of concentration on the creation process.
5.3 CARTON as a Smart Eyewear
In the survey, participants were asked to mark some characteristics, between 1 (bad) and 10 (good) regarding their experience on using CARTON as a smart eyewear device. Figure 10 shows a summary of their responses in a scale (1 to 5) consistent to other questions.
As Figure 10 illustrates, the users rated all characteristics of the demos above 3 except “Visibility” which shows some limits depending on the environment. We did not experiment outdoor with the participants, but we could already presume that it would be a limit as the display visibility of smartphones is difficult to adapt to very bright outdoor conditions. By using a slightly darker transparent plastic it would maybe decrease this issue, but it would also decrease a bit of its reachability by using a material harder to get. Noted that CARTON device has a peripheral see- through display, meaning the eyes have to focus on the display, which is quite different than other devices using stereoscopy technic such as Epson BT, Hololens and MetaVision. However, perhaps as a positive consequence of the precedent imperfect visibility, participants did not identify any issue due to eyes refocusing between the close display and long distance real world.
Despite that “Interactions” was the second best score, Voice Recognition could not be tested properly to navigate or interact in CARTON. This feature had been tested only in the tutorial and with the “Live Subtitle” but we could detect some issues. Sometime the device is not listening during 1 sec, which made users repeat. Only 45% of the participants felt that interacting with Head Gesture (tilting/nodding) was natural. This was mainly because the nod gesture had a critical issue. The forehead slope of each participant was relatively different and for some of them 15 degrees opposite tilt of the headset was not enough. The mobile (hence the axis using for detecting nodding) was always inclined and far from being horizontal as expected. It resulted that nod was not detected. Also one of the participants wanted a better tutorial, which seems quite important for unusual gesture. These results show that there is room for improvements with further research to this kind of head gesture recognition.
The ease to aim a target was judged to 4.0/5. Which is good enough to considering this kind of hardware compatible with AR technologies.
Interesting fact, even if it was not designed to, and not intended to be used in a production level (initially dedicated only to experimentation, research and development), half of the participants (6/12) were likely to wear CARTON in a museum. The main resistance of the others was because of the aesthetics. Finally, all the participants were enthusiastic, providing some new ideas, different use cases, how to improve it and showed a real interest. 66.7% of the participants said they would like to change and customize their CARTON to make it more accurate, robust and improve the aesthetics, or just to add colors and stickers. This result strengthens the importance to open and share the blueprints, guidelines and SDK related to this solution.
With all the data collected during the experiment, we already improved CARTON. For the hardware part there were a few minor upgrades to the blueprint (changed some lines to make it clearer and moved the rubber band’s location) and guide (instruction to choose a longer utility straps and the softest sponge possible which now fix trouble with the comfort part). Also, a video and a tutorial publicly accessible on YouTube (www.youtube.com/watch?v=3ww5lE8PVsc) and Instructables (a very famous DIY community website) have been published (www.instructables.com/id/Carton-DIY-Smart-Eyewear) to make the creation process easier and facilitates the distributed knowledge sharing [10, 16].
The Carton SDK for Android has also been updated. There is now a calibration feature for the Head Gesture Recognition. The tilt forehead is adapted to the wearer of the device, which makes the nod detection more accurate and efficient. Furthermore, there is another kind of head gesture implemented: head shaking (right/left), detected by counting a number of changing directions in an allotted time. The sample apps are in the process of publication on the Google Play Store, to make it easier for everyone to test it, before digging into it.
Despite the fact that CARTON is already usable, due to the lesson learned from the result, we will still slightly improve the comfort and visibility part mainly by trying new materials. Following a reviewer’s suggestion, we would like to add an aluminum foil between the cardboard and the sponge. We aim to measure the actual radiation to verify if this new material could protect the brain from the emitted smartphone’s radiation. Also, whereas that some effort was made taking into consideration CARTON’s accessibility for people with special needs (e.g., different ways to interact with CARTON), it is definitely not enough and we want to go further into this domain. By following guidelines to design accessible wearable technology, as proposed by Wentzel et al. , we would then be able to use it as a prototype for this area. It would be interesting to go deeper into gesture recognition research which could in addition to improve it, help to make it more accessible too. For instance, we could use the mirror module to add hand gesture recognition, a natural and intuitive way to interact with AR systems . We also hope to create a community around CARTON, which could for example develop an iOS SDK, Unity Package and sample apps with other AR technologies such as ARToolKit . We also may consider doing it ourselves. Also, despite the fact that we were against at the beginning, we consider now creating a 3D printable version of CARTON that we will publish openly on platforms such as Thingiverse. This motivation raised since we learned from the study that users may also want a plastic version, also community labs, which provide access to 3D printers, are getting more and more accessible and known by people. Furthermore, another conceivable approach would be to contact some professional maker (e.g., Dodocase), which then could provide a full kit with some preconceived part based on the free blueprint and guide. This solution may be interesting for two reasons: firstly, preconceived part should be more accurate than doing it manually, thus the conception will be quicker and the result more accurate and robust. Secondly, the constructor should be able to provide an even cheaper solution thanks to economies of scale, which would benefit for everyone by making the CARTON even more widely reachable. Finally, now that we can use CARTON as a smart eyewear, we aim to apply it to conduct research into collaborative multi-user and mobile/wearable computing.
To conclude, this paper presents an initiative, the CARTON project, to create a reachable smart eyewear device based on a DIY approach. Thus, this project allows end users to easily access another kind of multimedia (e.g., interface/interaction related to smart eyewear) just with a mobile phone. Moreover, this initiative includes an SDK to support the creation or the adaptation of existing Android apps for the CARTON device. We conduct a study to evaluate the effectiveness of the guidelines for building a CARTON device, while testing at the same time the usability of the proposed device through manipulation of provided sample apps. The results showed that all participants were able to create an effective CARTON with simple material and tools despite minor issues appeared due to heterogeneity of people physiognomy and skills. Thus, the contribution of this paper is double: (i) providing blueprints and guidelines to build a low-cost, open source and easy to do DIY smart eyewear devices and (ii) providing an SDK to support the development and the adaptation of mobile apps for eye-wearable.
By providing all of our works openly, not only our blueprint and code but also guides and full documentation, it makes much easier to reproduce experimentation, improvements, and to involve a community in producing a new version of CARTON. The fact of sharing is definitely not new and unique, but we still hope some other studies will follow this path, in this area or another.
Finding the appropriate balance between reachability and effectiveness was one of the hardest part related to this project. Is sacrificing reachability with such a material or tool worth it? This is particularly what the DIY community is very good at: finding new creative ways by combining multiple domains such as crafting, electronic, sewing ... a perfect match for mobile and wearable technology.
We would like to thank all the participants, and all publications support and staff, who wrote and provided helpful comments on previous versions of this document. As well authors gratefully acknowledge the grant from NSERCC (Natural Sciences and Engineering Research Council of Canada) and the company Black Artick.
 Ahmed, A., and Peralez P. 2014. Affordable Altered Perspectives: Making Augmented and Virtual Reality Technology Accessible. In Proc. IEEE GHTC, 603–8.
 ARToolKit SDK website: http://artoolkit.org/
 Bastien, J.M. C., and Scapin D. L. 1993. Ergonomic Criteria for the Evaluation of Human-Computer Interfaces. INRIA.
 Bimber O., Raskar R. 2003 Alternative Augmented Reality Approaches: Concepts, Techniques, and Applications. Eurographics.
 Buechley L., Rosner D. K., Paulos E., and Williams A. 2009. DIY for CHI: Methods, Communities, and Values of Reuse and Customization. Ext. Abstracts CHI 2009, ACM Press (2009), 4823–26.
 De Angeli, D., and O’Neill E. J. 2015. Development of an Inexpensive Augmented Reality (AR) Headset. In Proc. CHI 2015, ACM Press (2015), 971–76.
 Genaro Motti V., and Caine K. 2014. Understanding the Wearability of Head-Mounted Devices from a Human- Centered Perspective. In Proc. ISWC 2014, ACM Press (2014), 83–86.
 Google Cardboard website: http://google.com/cardboard
 Google, September Press Event 2015. YouTube video: https://www.youtube.com/watch?v=Jc-LEG0T_4c  Kuznetsov S., and Paulos E. 2010. Rise of the Expert Amateur: DIY Projects, Communities, and Cultures. In Proc. NordiCHI 2010, ACM Press (2010), 295–304.  Metavision device website: http://metavision.com/  Morency L.-P., and Darrell T. 2006. Head Gesture Recognition in Intelligent Interfaces: The Role of Context in Improving Recognition. In Proc. IUI 2006, ACM Press (2006) 32–38.  Naini, Farhad B. 2011. Facial Aesthetics: Concepts and Clinical Diagnosis. John Wiley & Sons.  Piumsomboon T., Clark A., Billinghurst M., and Cockburn A. 2013. User-Defined Gestures for Augmented Reality. Ext. Abstracts CHI 2013, ACM Press (2013).  Seebright device website: http://seebright.com/  Tanenbaum J. G., Williams A. M., Desjardins A., and Tanenbaum K. 2013. Democratizing Technology: Pleasure, Utility and Expressiveness in DIY and Maker Practice. In Proc. CHI 2013. ACM Press (2013), 2603–12.  Vuforia AR platform website: http://www.vuforia.com/  Wentzel J., Velleman E., and Geest T. 2016. Wearables for All: Development of Guidelines to Stimulate Accessible Wearable Technology Design. In W4A 2016.  Wille M., Adolph L., Grauel B., Wischniewski S., Theis S., and Alexander T. 2014. Prolonged Work with Head Mounted Displays. In Proc. ISWC 2014, ACM Press (2014), 221–24.  Yamauchi Y., Yokozawa M., Shinohara T., and Ishida T. 2000. Collaboration with Lean Media: How Open-Source Software Succeeds. In Proc. CSCW 2000. ACM Press (2000), 329–38
Damien Brun LICEF Research Center, Télé-Université du Québec, Montréal, QC, Canada - email@example.com Susan M. Ferreira LICEF Research Center, Télé-Université du Québec, Montréal, QC, Canada - firstname.lastname@example.org Charles Gouin- Vallerand LICEF Research Center, Télé-Université du Québec, Montréal, QC, Canada - email@example.com Sébastien George UBL, Université du Maine, EA 4023, LIUM, 72085 Le Mans, France -sebastien.george@univ- lemans.fr