Amebous Labs https://amebouslabs.com Tue, 05 Jul 2022 15:48:50 +0000 en-US hourly 1 https://wordpress.org/?v=6.5.5 https://amebouslabs.com/wp-content/uploads/2020/06/cropped-Controller_Full-Color-32x32.png Amebous Labs https://amebouslabs.com 32 32 UNITY UI REMOTE INPUT https://amebouslabs.com/unity-ui-remote-input/ Wed, 29 Jun 2022 09:30:00 +0000 https://amebouslabs.com/?p=4551

UNITY UI REMOTE INPUT

Pierce McBride

Written by Pierce McBride

June 29, 2022

Throughout many VR (virtual reality) games and experiences, it’s sometimes necessary to have the player interact with a UI (user interface). In Unity, these are commonly world-space canvases where the player usually has some type of laser pointer, or they can directly touch the UI with their virtual hands. Like me, you may have in the past wanted to implement a system and ran into the problem of how can we determine what UI elements the player is pointing at, and how can they interact with those elements? This article will be a bit technical and assumes some intermediate knowledge of C#. I’ll also say now that the code in this article is heavily based on the XR Interaction Toolkit provided by Unity. I’ll point out where I took code from that package as it comes up, but the intent is for this code to work entirely independently from any specific hardware. If you want to interact with world-space UI with a world-space pointer and you know a little C#, read on.

Figure 1 – Cities VR (Fast Travel Games, 2022) 

Figure 2 – Resident Evil 4 VR (Armature Studio, 2021)  

Figure 3 – The Walking Dead: Saints and Sinners (Skydance Interactive, 2020) 

It may be tempting to say, “Well, I can simply use Physics.Raycast to find out what the player is pointing at,” and unfortunately, you would be wrong like I was. That’s a little glib, but in essence, there are two problems with that approach:    

  1. Unity uses a combination of the EventSystem, GraphicsRaycasters, and screen positions to determine UI input, so by default, Physics.Raycasts would entirely miss the UI with its default configuration.

  2. Even if you attach colliders and find the GameObject hit with a UI component, you’d need to replicate a lot of code that Unity provides for free. Pointers don’t just click whatever component they’re over; they can also drag and scroll.  

In order to explain the former problem, it’s best to make sure we both understand how the existing EventSystem works. By default, when a player moves a pointer (like a mouse) around or presses some input that we have mapped to UI input, like select or move, Unity picks up this input from an InputModule component attached to the same GameObject as the EventSystem. This is, by default, either the Standalone Input Module or the Input System UI Input Module. Both work the same way, but they use different input APIs.   

Each frame, the EventSystem calls a method in the input module named Process(). In the standard implementations, the input module gets a reference to all enabled BaseRaycaster components from a RaycasterManager static class. By default, these are GraphicsRaycastersfor most games. For each of those raycasters, the input module called Raycast(PointerEventData eventData, List<RaycastResult> resultAppendList) takes a PointerEventData object and a list to append new graphics to. The raycaster sorts all the graphic’s objects on its canvas by a hierarchy that lines up with a raycast from the pointer screen position to the canvas and appends those graphics objects to the list. The input module then takes that list and processes any events like selecting, dragging, etc. 

Figure 4 Event System Explanation Diagram 

Now, that whole process essentially skips anything related to physics, hence why you won’t be able to simply raycast to determine what the player is pointing at without additional setup. Though, in my opinion, the second issue is the more problematic one because you’re skipping all this basic functionality, and the best programs don’t need systems to be written twice. We don’t want to write duplicative code and make our jobs harder, which begs the question, what is the alternative? The answer is to build our own Raycaster, EventData, and InputModule. The beauty of this approach is it lets our Input and our UI be weakly coupled but highly cohesive. In other words, your UI doesn’t need to know or care that it’s running in a VR game or is receiving input in a non-standard way.  

So how will these objects fit together? Instead of the process described above, we’ll replace the input module with a RemoteInputModule, each raycaster with a RemoteInputRaycaster, create a new kind of pointer data called RemoteInputPointerData, and finally make an interface for an IRemoteInputProvider. You’ll construct the interface yourself to fit the needs of your project, and its job will be to register itself with the input module, update a cached copy of RemoteInputPointerData each frame with its current transform position rotation, and set the Select state which we’ll use to actually select UI elements. 

Each time the EventSystem calls Process() on our RemoteInputModule we’ll refer to a collection of registered IRemoteInputProvider and retrieve the corresponding RemoteInputEventData. InputProviderSet is a Hashset for fast lookup and no duplicate items. Lookup is a Dictionary so we can easily associate the providers with a cached event data. We cache event data so we can properly process UI interactions that take place over multiple frames, like drags. We also presume that each event data has already been updated with a new position, rotation, and selections state. Again, this is your job as a developer to define how that happens, but the RemoteInput package will come with one out of the box and a sample implementation that uses keyboard input that you can review. 

Next, we need all the RemoteInputRaycaster. We could use Unity’s RaycastManager, iterate over it, and call Raycast() on all BaseRaycaster that we can cast to our inherited type, but to me, this sounds slightly less efficient than we want. We aren’t implementing any code that could support normal pointer events from a mouse or touch, so there’s not really any point in iterating over a list that could contain many elements we have to skip. Instead, I added a static HashSet to RemoteInputRaycaster that it registers to on Enable/Disable, just like the normal RaycastManager. But in this case, we can be sure each item is the right type, and we only iterate over items that we can use. We call Raycast(), which creates a new ray from the provider’s given position, rotation, and max length. It sorts all its canvas graphics components just like the regular raycaster. 

Last, we take the list that all RemoteInputRaycaster have appended and process UI events. SelectDelta is more critical than SelectDown. Technically our input module only needs to know the delta state because most events are driven when select is pressed or released only. In fact, RemoteInputModule will set the value of SelectDelta to NoChange after it’s processed each event data. That way, it’s only ever pressed or released for exactly one provider.

Figure 5 Remote Input Process Event Diagram 

To get started on the actual implementation walk through the snippets and refer to the sample and the package repo. You are welcome to use the package for commercial and personal uses, just follow the requirements of the license. It’s configured as a Unity package, so as long as you add the package as a git repo to your project, it will download itself to your project.

FuturusDev/FuturusRemoteInput: A Unity Package for interacting world-space UI from a world-space remote, by Futurus (github.com) 

For you, the most important code to review would be our RemoteInputSender because you should only need to replace your input module and all GraphicsRaycasters for this to work. Thankfully, beyond implementing the required properties on the interface, the minimum setup is quite simple.

void OnEnable() 
{ 
    ValidateProvider(); 
    ValidatePresentation(); 
} 
void OnDisable() 
{ 
     _cachedRemoteInputModule?.Deregister(this); 
}
void Update() 
{ 
     if (!ValidateProvider()) 
         return; 
     _cachedEventData = _cachedRemoteInputModule.GetRemoteInputEventData(this); 
     UpdateLine(_cachedEventData.LastRaycastResult); 
     _cachedEventData.UpdateFromRemote(); 
     SelectDelta = ButtonDeltaState.NoChange; 
     if (ValidatePresentation()) 
        UpdatePresentation(); 
}
On enable, we validate our provider and presentation. Validating our provider just means caching a reference to the RemoteInputModule through the singleton reference EventSystem.current and registering ourselves with SetRegistration(this, true). Because we store registered providers in a Hashset, you can call this each time, and no duplicate entries will be added. Validating our presentation means updating the properties on our LineRenderer if it’s been set in the inspector.
bool ValidatePresentation() 
{ 
    _lineRenderer = (_lineRenderer != null) ? _lineRenderer : GetComponent 	&lt;linerenderer&gt;(); 
    if (_lineRenderer == null) 
        return false; 
 
    _lineRenderer.widthMultiplier = _lineWidth; 
    _lineRenderer.widthCurve = _widthCurve; 
    _lineRenderer.colorGradient = _gradient; 
    _lineRenderer.positionCount = _points.Length; 
    if (_cursorSprite != null &amp;amp;&amp;amp; _cursorMat == null) 
    { 
        _cursorMat = new Material(Shader.Find("Sprites/Default")); 
        _cursorMat.SetTexture("_MainTex", _cursorSprite.texture); 
        _cursorMat.renderQueue = 4000; // Set renderqueue so it renders above existing UI 
        // There's a known issue here where this cursor does NOT render above dropdown components.  
        // it's due to something in how dropdowns create a new canvas and manipulate its sorting order, 
        // and since we draw our cursor directly to the Graphics API we can't use CPU-based sorting layers 
        // if you have this issue, I recommend drawing the cursor as an unlit mesh instead 
        if (_cursorMesh == null) 
            _cursorMesh = Resources.GetBuiltinResource&amp;lt;Mesh&amp;gt;("Quad.fbx"); 
    } 
 

    return true; 
}
RemoteInputSender will still work if no LineRenderer is added, but if you add one it’ll update it via the properties on the sender each frame. As an extra treat, I also added a simple “cursor” implementation that creates a cached quad, assigns a sprite material to it that uses a provided sprite and aligns it per-frame to the endpoint of the remote input ray. Take note of Resources.GetBuiltinResource("Quad.fbx"). This line gets the same file that’s used if you hit Create -> 3D Object -> Quad and works at runtime because it’s a part of the Resources API. Refer to this link for more details and other strings you can use to find the other built-in resources. 

The two most important lines in Update are _cachedEventdata.UpdateFromRemote() and SelectDelta = ButtonDeltaState.NoChange. The first line will automatically set all properties in the event data object based on the properties implemented from the provider interface. As long as you call this method per frame and you write your properties correctly the provider will work. The second resets the SelectDelta property, which is used to determine if the remote provider just pressed or released select. The UI is built around input down and up events, so we need to mimic that behavior and make sure if SelectDelta changes in a frame, it only remains pressed or released for exactly 1 frame.

void UpdateLine(RaycastResult result) 
{ 
    _hasHit = result.isValid; 
    if (result.isValid) 
    { 
        _endpoint = result.worldPosition; 
        // result.worldNormal doesn't work properly, seems to always have the normal face directly up 
        // instead, we calculate the normal via the inverse of the forward vector on what we hit. Unity UI elements 
        // by default face away from the user, so we use that assumption to find the true "normal" 
        // If you use a curved UI canvas this likely will not work  
        _endpointNormal = result.gameObject.transform.forward * -1; 
    } 
    else 
    { 
        _endpoint = transform.position + transform.forward * _maxLength; 
        _endpointNormal = (transform.position - _endpoint).normalized; 
    } 
    _points[0] = transform.position; 
    _points[_points.Length - 1] = _endpoint; 
} 
void UpdatePresentation() 
{ 
    _lineRenderer.enabled = ValidRaycastHit; 
    if (!ValidRaycastHit) 
        return;  

    _lineRenderer.SetPositions(_points); 
    if (_cursorMesh != null &amp;amp;amp;&amp;amp;amp; _cursorMat != null) 
    { 
        _cursorMat.color = _gradient.Evaluate(1); 
        var matrix = Matrix4x4.TRS(_points[1], Quaternion.Euler(_endpointNormal), Vector3.one * _cursorScale); 
        Graphics.DrawMesh(_cursorMesh, matrix, _cursorMat, 0); 
    } 
} 
These last two methods simply update the “presentation” of the RemoteInputSender.  Because the LastRaycastResult is cached in the event data, we can use it to update our presentation efficiently. In most cases we likely just want to render a line from the provider to the UI that’s being used, so we use an array of Vector3 that’s of length 2 and update the first and last position with the provider position and raycast endpoint that the raycaster updated. There’s an issue currently where the world normal isn’t set properly, and since we use it with the cursor, we set it ourselves with the start and end point instead. When we update presentation, we set the positions of the line renderer, and we draw the cursor if we can. The cursor is drawn directly using the Graphics API, so if there’s no hit, it won’t be drawn and has no additional GameObject or component overhead. 

I would encourage you to read the implementation in the repo, but at a high level, the diagrams and explanations above should be enough to use the system I’ve put together. I acknowledge that portions of the code use methods written for Unity’s XR Interaction Toolkit, but I prefer this implementation because it is more flexible and has no additional dependencies. I would expect most developers who need to solve this problem are working in XR, but if you need to use world space UIs and remotes in a non-XR game, this implementation would work just as well.

Figure 6 Non-VR Demo 

Figure 7 VR Demo 

Although this is certainly not the simplest implementation of VR UI interactions, it is the most stable and robust implementation I’ve ever used, and once it’s written, it shouldn’t need to be modified. We at Futurus have a ready-made package on Github that you can find linked below. You can insert this into your projects today and try it out, or you can peruse the repo yourself. There are a few other helper classes and types that I didn’t cover for the sake of length, but they’re present in the repo if you want to see them.  

FuturusDev/FuturusRemoteInput: A Unity Package for interacting world-space UI from a world-space remote, by Futurus (github.com) 

Cheers, and good luck building your game ✌️ 

]]>
GDC 2022: DAY 4 https://amebouslabs.com/gdc-2022-day-4/ Fri, 25 Mar 2022 14:16:04 +0000 https://amebouslabs.com/?p=4454

GDC 2022: DAY 4

Annie Eaton

Written by Nick Foster

March 25, 2022

Game Developers Conference (GDC) 2022 may be nearing its end, but things are definitely not winding down! The fourth day of GDC was packed with inspiring hardware demos and even more game-changing presentations from some of gaming’s most creative minds. Read on for our favorite parts of GDC day four and a huge Loam Sandbox surprise!

Chan’s Picks:

Building Next-Gen Games for PlayStation VR2 with Unity

On day four, I went to an interesting talk presented by Unity regarding its support for PlaystationVR2 games. It highlighted new PSVR2 features like eye-tracking and Foveated Rendering while detailing how Unity will implement them. While much of the talk focused on PSVR2’s exclusive features, it also explored Unity’s broader features and what we can expect regarding game development on the system moving forward. 

While URP and Single-Pass rendering have been talked about ad nauseam, I find explanations for some of their advanced uses to be fascinating. For example, I know that you can make a custom rendering pipeline for URP, but I didn’t consider that I could have a pipeline focused solely on UI which can be run separately from everything else.

Tilt Five uses AR-enabled glasses and a unique controller to breathe life into our favorite table-top board games.

I also walked around the GDC Expo Hall and saw innovative games with alternative control schemes. The artistry and creativity on display really encouraged me to find new and novel ways to play.

Pierce’s Picks:

The MAW: Safely Multithreading the Deterministic Gameplay of ‘Age of Empires IV’

Joel Pritchett (Technical Director of Age of Empires Franchise, Microsoft) details the multithreading, debugging, and verification tools built into the game’s engine.

Age of Empires is a large-scale game featuring many, many different units and entities. Whether you are running the game on a single-core CPU or a massive gaming PC, the sim game must always produce a similar experience, so lots of thought went into designing the game’s engine to adhere to this requirement. The game’s developers parallelize units by grouping them into islands — these are groups of units that need to know about one another but not about other islands.

RTSs used deterministic input syncing, and it usually worked fine: the peers agree on input and return the same result. The issue with multithreading, however, is that it was possible for threads to start or stop at different times in different machines, making it potentially non-deterministic. The developers resolved this issue by recording read/write records on data in the engine; if there were any conflicts between peers, they could detect those errors. With this in place, developers could just write code, and whenever they wrote code that would create desyncs, they were automatically and immediately caught.

Amy’s Picks:

Understanding Developer Challenges with Localization, Player Support, and More

Annie, Alexander, Xavier, and Ninel share actionable advice on how to best localize your game or application in order to create an experience that many can enjoy.

My pick isn’t from day four but rather the end of day three. Annie, the executive producer at Amebous Labs, joined the TransPerfect panel with host Alexander Fletcher (Global Director, TransPerfect Gaming Solutions) and co-panelists Xavier Marot (Director of Production, Focus Home Interactive) and Ninel Anderson (CEO, Devoted Studios).

Together, they shared an insightful discussion regarding how developers can create a seamless experience for players from all over the world. Ninel suggested tapping into the community of players that speak the language you’ve translated your game into and trading them game keys for feedback. Taking this step will not only help the developers, but it will make the players feel heard as they offer advice on how to appropriately translate content for their audience.

Annie shared ways to avoid using text in virtual reality, as it is inherently difficult to read large amounts of it in virtual reality. She suggested using iconography and visual diagrams to assist the player in successfully maneuvering throughout a VR game.

Annie’s Picks:

Hardware Demo: Tilt Five

Amy and Pierce play an AR board game on the Tilt Five.

We got to have a private demo of Tilt Five, an augmented reality wearable for tabletop games. The weight of this device was lighter than other HMDs and the field of view is impressive (especially for the weight). I think we’ll be putting an order in for a couple of these so that we can experiment with creating content for this device!

We just added brand new critters, crops, and decorations to the magical world of Loam Sandbox! For more information on Loam Sandbox’s biggest update yet, drop by the #release-updates channel of our discord server.

]]>
GDC 2022: DAY 3 https://amebouslabs.com/gdc-2022-day-3/ Thu, 24 Mar 2022 15:49:58 +0000 https://amebouslabs.com/?p=4411

GDC 2022: DAY 3

Annie Eaton

Written by Nick Foster

March 24, 2022
The Game Developers Conference (GDC) is in full gear! Throughout day three of GDC, we celebrated Atlanta’s game development community, attended insightful presentations, and explored new games and technology featured on the Expo floor. Continue reading to see all of our favorite moments from the third day of GDC 2022.
Pierce’s Picks:

The Fine Line Between Difficult and Impossible: Adventures in VR Development

Devin Reimer (CEOwl, Owlchemy Labs) and Andrew Eiche (CTOwl & Cable Slinger, Owlchemy Labs) describe the difference between ‘impossible’ development problems and ‘very difficult’ ones.

Problem-solving and critical thinking are two major aspects of game development. Getting your game to look and run how you envisioned on various different hardware systems is difficult and, at times, borderline impossible. In this presentation led by Owlchemy Labs, the game development studio behind the hit VR game Job Simulator, Devin Reimer and Andrew Eiche argue that the best development rides the fine line between an impossible problem and a daunting challenge.

The Owlchemy team explained how something impossible feels the same as something very difficult, but impossibility can only be factored in when looking at the overall project holistically. Each individual issue may be difficult, but when you have too many difficult problems, then the entire project becomes impossible. During the development of Cosmonious High, Owlchemy Labs realized the importance of being flexible with your game’s requirements — the size and scope of Owlchemy’s original vision for the game had to be scaled back in order to effectively optimize the game.

Amy’s Picks:

Three Opportunities to Improve Character Design in the Games Industry

Jessica Tompkins (Senior Researcher, Electronic Arts) shares historical trends in game character design.

The expo show floor opened up today with so much to see and do, but I made it a point to attend Jessica Tompkins’ talk on character design. She shared her findings and ideas regarding how developers can be agents of change at their game studios and improve representation in video games. 

She suggested focusing on motivation driven design rather than designing for demographic parameters, going beyond being an ally and actually pushing for change by acting and advocating, and, finally, opening up the conversation while future game industry professionals are in school — it is not a conversation that higher education should be timid about having. Taking these steps will help to push the industry forward and be more welcoming of underrepresented gamers and game developers.

Annie’s Picks:

Hardware Demo: Speech Graphic’s Facial Animation Tool

Chan and I got to demo Speech Graphics’ audio-driven facial animation tool. It was fantastic seeing how their product creates extraordinarily realistic face and mouth movements from just a single audio file. Tools like these will be monumental in expediting production pipelines, so I can’t wait to try this for some of our future VR projects.

Chan’s Picks:

Serious Games Roundtable

Matthew Lee (Chair, IGDA Serious Games SIG) chats with developers about various styles of game design in a roundtable discussion.

I had the absolute privilege of attending the Serious Games Roundtable on the third day of GDC. I admit I was not familiar with the roundtable format when I walked into the room, so I was very surprised to see a bunch of tables lined up in a circle. Each developer took a seat at the table and a live mic was passed around so that everyone who wanted to talk had an opportunity to do so. 

As a lifetime lover and admirer of both serious and educational game design, it was interesting to hear insights and thoughts from fellow developers. While I recognized a few terms tossed around during the discussion, I learned several new phrases and ideas that I absolutely adore and would love to start using regularly.

Game Economies: An Economist’s Perspective

I also attended a talk regarding in-game economies presented by Christopher Smith, the chief economist at Metanomic Ltd. He explained economic theories and concepts that should influence your game design. While most of it went over my head, I thought the talk was incredibly informative, especially considering our studio is currently supporting Loam Sandbox with post-launch content.

As a lifelong admirer of both serious and educational game design, it was interesting to hear insights and thoughts from fellow developers. While I recognized a few terms tossed around during the discussion, I learned several new phrases and ideas that I absolutely adore and would love to start using regularly.

Honorable Mention:

Understanding Developer Challenges with Localization, Player Support, and More

Annie shares VR game development problems and the best ways to tackle them.

During day three of GDC, Annie also joined Alexander Fletcher (Global Director, TransPerfect Gaming Solutions), Xavier Marot (Director of Production, Focus Home Interactive), and Ninel Anderson (CEO, Devoted Studios) on the GDC Industry Stage for a panel discussion exploring real-life game development issues and scenarios.

New crops, critters, and decor are coming soon to the magical world of Loam! We’re showcasing Loam Sandbox‘s upcoming content drop in the Transperfect Gaming Solutions booth (S1055) on Thursday, March 24 from 10 AM – 12 PM. If you’re at GDC, please stop by to preview the latest additions to the magical world of Loam—see you there!

]]>
GDC 2022: DAY 2 https://amebouslabs.com/gdc-2022-day-2/ Wed, 23 Mar 2022 14:56:37 +0000 https://amebouslabs.com/?p=4369

GDC 2022: DAY 2

Annie Eaton

Written by Nick Foster

March 23, 2022
The Game Developers Conference (GDC) is ramping up! Day two of the event presented our team with opportunities to use new hardware and learn new tips and tricks for mitigating user motion sickness and designing in-game environments. With several members of the Amebous Labs team attending the conference in San Francisco, here were some of our favorite moments from the second day of GDC 2022.
Amy’s Picks:

Art Direction Summit: Building Night City

Kacper Niepokólczycki (Lead Environment Artist, CD Projekt Red) explains the importance of emotion when designing environments for your game.

My favorite session of day two was presented by the lead environment artist at CD Projekt Red, Kacper Niepokólcyzcki; his talk was centered around the creation of Cyberpunk 2077‘s environments. My main takeaways were to think about how you can evoke emotions through the design of the environment from the very start. 

Although it may lack color and feature very limited detail, the earliest stages of your environment design should still make you feel something. The following steps of the design process should continue to strengthen that feeling you want the player to experience while playing your game. If it doesn’t, then you will need to re-iterate until you hit upon an environment better aligned with your game’s core pillars. When you are making design decisions, you should always stop and ask yourself (or your team) “WHY?” If the answer is vague or unclear, then the concept needs to be worked on.

Chan’s Picks:

Future Realities Summit: How NOT to Build a VR Arcade Game

Michael Bridgman (Co-founder & CTO, MajorMega) shares development tips for mitigating user motion sickness in VR experiences.

On my second day of GDC, I went to an awesome panel covering VR arcade games: the dos, don’ts, trials, and tribulations. Michael Bridgman, the developer hosting the presentation, has made incredible VR experiences, and he discussed a few tricks to help mitigate major causes of motion sickness in virtual reality. While some of it isn’t applicable to the kind of development that we do at Amebous Labs, I think he shared many interesting things worth looking into.

Day two also presented me with a chance to use the HTC Flow and the HTC Focus. The XR industry often focuses on the Oculus Quest and to a lesser extent the Vive Index, so it was very lovely to use some alternative headsets. I want to give more attention to the HTC Flow which is a lightweight VR system that takes advantage of your phone’s processing power to reduce the overall size of the headset. Five minutes with it on and I was already trying to formulate ideas for how we might use it (and how I would convince Annie to buy it…)

Annie’s Picks:

Meeting HTC Vive’s Developer Community

Annie tries out the HTC Focus 3’s new wrist trackers.

It was great to meet more people in the HTC Vive developer community! On day two, I got to see the new Wrist Tracker device for the Vive Focus 3. It was great to see how this hardware has enhanced hand tracking and reduced issues with occlusion.

Pierce’s Picks:

Machine Learning Summit: Walk Lizzie, Walk! Emergent Physics-Based Animation through Reinforcement Learning

Jorge del Val Santos (Senior Research Engineer, Embark Studios) highlights a new method for animating digital assets—machine learning!

The traditional way of bringing digital assets to life consists of first rigging and then animating them, but machine learning agents (ML-Agents) present a new, physics-based option. Using a reinforcement learning approach to automatic animation, developers can bring life to creatures and critters with the click of a button — no animators involved. With that said, a fair few challenges accompany the benefits of this technique, so it may not always be the best fit for your game.

One of the greatest challenges to this technique is reward structure; because ML-Agents will be rewarded no matter how they move, they might end up learning to walk in unintended ways. For example, mixups in reward structures can cause a creature to walk using an arm and a leg because there was nothing in place disfavoring that. Another weakness of this physics-based model of animation is authorship or, in other words, your ability to bring character to your creations. At the moment, there’s no clear way of training your monsters to be scary or move intimidatingly.

If you’re in San Francisco for GDC, we’d love to say hi! We’re showcasing Loam Sandbox at the Transperfect Gaming Solutions booth (S1055) on Wednesday, March 23 from 2 – 4 PM PDT and on Thursday, March 24 from 10 AM – 12 PM, so please come stop by. We also look forward to celebrating Atlanta’s talented community game developers at GDC alongside the Metro Atlanta Chamber and the Atlanta Esports Alliance on Wednesday evening; register here to join us at Local Kitchen for a fun happy hour reception.

]]>
GDC 2022: DAY 1 https://amebouslabs.com/gdc-2022-day-1/ Tue, 22 Mar 2022 14:49:09 +0000 https://amebouslabs.com/?p=4294

GDC 2022: DAY 1

Annie Eaton

Written by Nick Foster

March 22, 2022
The Game Developers Conference (GDC) has finally begun! Every year, the event seeks to bring the gaming industry together for a week of education, inspiration, and networking in San Francisco, California. This year, several members of the Amebous Labs team flew out to share ideas, meet fellow gamedevs, and even showcase Loam Sandbox.

 

Filled with insightful summit discussions and presentations tackling a variety of topics, day one of GDC had something for everyone; read on to discover some of our favorite moments.
Annie’s Picks:

Future Realities Summit: How NASA Has Translated Aerospace Research into Biofeedback Game Experiences

John E. Muñoz (Game Designer & Postdoctoral Fellow, Personal) and Alan Pope (Distinguished Research Associate, NASA) explore gaming’s adoption of NASA’s biofeedback research.

It was fascinating to see how NASA’s advanced biofeedback device research has been integrated into games. As the technology advanced, biofeedback was used to do psychophysiological modeling in virtual reality. The ability to track one’s heart rate and frontal alpha has produced more effective conflict de-escalation training, leveraging biocybernetic adaptation. Applications utilizing this technology have also helped researchers further understand the brain functioning of individuals with autism and ADHD.

Animation Summit: Animation and Customization in ‘The Sims 4’

Yusun Chung (Lead Animator, Electronic Arts) analyzes aesthetic considerations and implementation workflows necessary to create animations for customizable characters, clothing, and interactable objects.

Pierce and I attended a great session covering the animation and customization in ‘The Sims 4’. This was especially of interest due to our future goals of implementing customization in Loam Sandbox. It was helpful to learn about the importance of object variety when an asset type must feature animations or interactivity. By setting parameters for artists, an asset’s core interactivity can stay the same for variants while the outer shell can be as creative as the artist wants.

Chan’s Picks:

Future Realities Summit: Up Close & Virtual – The Power of Live VR Actors

Alex Coulombe (Creative Director, Agile Lens: Immersive Design) details the techniques, tools, and technology he uses to create compelling VR shows.

I attended a panel called Up Close & Virtual presented by the CEO of Heavenue, Alex Coulombe. He discussed his company’s experience creating a live VR theater performance. In his performance of the Christmas Carol, various techniques and technologies combine to create a one-man show for audiences to enjoy in VR. I was absolutely impressed by the idea and I wouldn’t mind exploring more of it in the future. We could be seeing the creation of a new type of theater that is solely unique to VR!

Honestly, talking to fellow developers throughout the day has been among the real treats of GDC. I’ve met several VR developers, and it’s interesting to hear their opinions and thoughts on game development and existing experiences for VR platforms such as PSVR, Oculus/Meta, and Vive.

Pierce’s Pick:

Future Realities Summit: Hands Best: How AR & VR Devs Can Make the Most of Hand Tracking

Brian Schwab (Director of Interaction & Creative Play Lab, LEGO Group) and John Blau (Game Designer, Schell Games) share their philosophies for designing hand tracking in VR.

VR game development forces you to reframe the way you think about your hands and the many ways you use them to interact. It was interesting to hear John and Brian refer to hands as systems through which we attain information about the things around us by reaching out and touching or gauging distance.

There are so many different factors to keep in mind when designing and implementing hand tracking for your VR apps such as user comfort, cultural differences, and the tracking complexity of your experience. However, as a general rule of thumb, developers should first identify the type of experience they want to create before they begin designing any hand tracking for it. Once you’ve realized this, choosing a combination of interaction models to match your goals can become a less overwhelming process.

Amy’s Pick:

Visual Effects Summit: How to (Not) Create Textures for VFX

Simon Trümpler emphasizes how using photos and third-party assets saves time when making textures.

Many artists feel like creating new textures for visual effects using existing photos and third-party assets is cheating, but Simon dispelled these common beliefs and showed us how we can take a base asset and customize it in a way that makes it our own. His advice was to focus on getting assets to the “good enough” stage because, if you find additional time later on in the game development process, then you can fine-tune them to have increasingly customized looks.

One of my favorite tricks he demonstrated was taking a picture of a forest, flipping it upside down, bumping up the contrast, then combining it with other water effects he had already created. The result was a waterfall that appeared to have ripples and movement. Had he painted each line by hand, it would have taken significantly longer to create. He wrapped up the talk by encouraging artists to share their expertise and learnings on social media using #VFXTexture to streamline the texture creation process for fellow artists.

If you’re in San Francisco for GDC, we’d love to say hi! We’re showcasing Loam Sandbox at the Transperfect Gaming Solutions booth (S1055) on Wednesday, March 23 from 2 – 4 PM PDT and on Thursday, March 24 from 10 AM – 12 PM, so please come stop by. We also look forward to celebrating Atlanta’s talented community game developers at GDC alongside the Metro Atlanta Chamber and the Atlanta Esports Alliance on Wednesday evening; register here to join us at Local Kitchen for a fun happy hour reception.

]]>
LOAM SANDBOX POST-LAUNCH ROADMAP https://amebouslabs.com/loam-sandbox-post-launch-roadmap/ Tue, 22 Feb 2022 09:29:00 +0000 https://amebouslabs.com/?p=4240

LOAM SANDBOX POST-LAUNCH ROADMAP

Annie Eaton

Written by Nick Foster

February 22, 2022
In the Spring of 2020, we envisioned a virtual world in which players would have the freedom to design, create, and nurture their unique gardens year-round—an experience that would eventually become Loam Sandbox. Now, two years and an App Lab launch later, our focus remains on developing Loam’s magical world.
Loam Sandbox is not finished growing.
The team is excited to see our garden sim game concept finally come to life, but, even now, Loam Sandbox is not finished growing. It feels like every day, our developers, artists, and producers think up another critter, plant, or piece of décor that would be a great addition to Loam; though certain items will undoubtedly take some time to implement, we are happy to share that we will continue to support, refine, and expand upon the world of Loam Sandbox through future updates and fixes.

 

We want player feedback, above all else, to be at the epicenter of every change we make to Loam. In other words, you have the power to influence Loam’s post-launch trajectory by sharing your thoughts and desires with us via the #feedback channel of Loam’s Discord server.
Because players are so important to Loam Sandbox’s post-launch support, we want to be as transparent as possible regarding what our team is working on and any updates we’re planning to release. That’s why we recently created a public Trello board detailing Loam’s progress in real-time. While the cards listed throughout this roadmap are subject to change and not completely guaranteed to make it into the game, Trello will be a helpful way for our developers to keep you in the loop as Loam continues to grow and evolve.

When you share your feedback, suggestions, and bugs with us in Discord, they’ll eventually make their way onto the board’s “Player Feedback & Suggestions” list in the form of a brand-new card. Depending on their priority, these cards will move to the “To Do” list, then to” In Progress,” before finally earning a spot on the board’s “Done” list. As we wrap things up and roll out Loam Sandbox updates, we’ll also share what has changed via our Discord’s #release-updates channel.

Loam’s public roadmap Trello board also gives you the power to voice your opinion; let us know which changes you want to see made to Loam Sandbox next by voting for your favorite cards. As the list of suggestions continues to grow, this will help our team prioritize various updates. As of writing, we’re currently working toward implementing new décor items and making several quality of life changes to the communications console that will make it more user-friendly.

Click here to visit the Trello board and track Loam’s development in real-time.

Did you know our team occasionally plays Loam Sandbox live on Twitch and YouTube? Check out our favorite plants and décor, learn new gardening techniques, and even meet a few of Loam Sandbox’s developers, artists, and producers by watching full replays of our live-streams.

]]>
LOAM SANDBOX IS OUT NOW ON QUEST APP LAB https://amebouslabs.com/loam-sandbox-is-out-now-on-quest-app-lab/ Thu, 27 Jan 2022 21:29:00 +0000 https://amebouslabs.com/?p=4200

LOAM SANDBOX IS OUT NOW ON QUEST APP LAB

Annie Eaton

Written by Nick Foster

January 27, 2022

We are proud to officially announce the release of Amebous Labs’ debut title, Loam Sandbox! Our early access garden sim game aims to provide players with a comforting, creative virtual reality experience.

In the magical world of Loam Sandbox, every player begins with a near-empty garden space they can personalize to their liking by harvesting and selling flowers, herbs, fruits, and vegetables to purchase more seeds and new decor. As players transform plots into the gardens of their dreams, they will attract cute critters, adding unique life to every garden design!
Personalize your garden by growing plants, buying decor, and attracting critters.
“Our talented team of developers, artists, designers, and creatives have come together to create something truly magical. We wanted to produce a VR title that allowed players to escape into their own world where they could design, create, and nurture a garden year-round,” said Executive Producer Annie Eaton. “We’ve overcome many challenges to develop an experience that gives players a fully customizable space, built for mobile VR first. We hope that people of all backgrounds will enjoy bringing their own touch to the land of Loam.”

From the start of development in early 2020, Loam’s goal has been to give players the ability to enter a comforting world where they can relax and express their creativity. However, just like the seeds available within the game, Loam is constantly growing; Amebous Labs is requesting player feedback to ensure that all future updates and downloadable content are as positive as possible. Let us know what you want to see more of by commenting in the #feedback channel of our Discord server. And check out our public roadmap to see what features we are working toward adding to Loam Sandbox next.

“Our team put forth a great deal of effort and care into making a virtual reality game that is as beautiful as it is relaxing,” said Producer Amy Stout. “As the game continues to evolve, we hope to add special touches that provide more avenues for players to showcase their individuality.”

It’s time to dust off your gardening gloves—Loam Sandbox is available now exclusively on the Meta Quest platform for $14.99. Download Loam Sandbox for your Meta Quest or Meta Quest 2 headset!

Annie Eaton

Written by Nick Foster

]]>
FINALIZATION OPTIMIZATION: PUBLISHING QUEST VR APPS AND GAMES https://amebouslabs.com/finalization-optimization-publishing-your-vr-apps-and-games/ Wed, 01 Dec 2021 16:30:24 +0000 https://amebouslabs.com/?p=3970

FINALIZATION OPTIMIZATION: PUBLISHING QUEST VR APPS AND GAMES

Annie Eaton

Written by Nick Foster

December 1, 2021

We’re closer than ever to Loam Sandbox’s early access release via App Lab. For nearly two years, our team has been working tirelessly to bring the vision of a relaxing virtual reality garden sim game to life. Much of this time was spent designing the game, but a surprising amount went toward optimizing Loam to adhere to App Lab’s standards and requirements. Developing and optimizing a VR game has provided an invaluable opportunity for us to familiarize ourselves with the Quest Virtual Reality Check (VRC) guidelines, an official list of app requirements and recommendations set forth by Oculus (soon to be officially rebranded as Meta). Throughout this blog, we will share the tricks we’ve learned in the hopes that they will help you get your games and apps running on Oculus Quest headsets.

Developers should determine their app’s ideal distribution channel as early as possible.

As you prepare your work for release, the distribution channel you will use to house your game is a significant decision you’ll need to consider as early as possible. Your decision will directly determine how much time and effort you’ll need to spend optimizing your game to run in accordance with various VRCs. Developers tasked with getting their work published and accessible via Quest headsets are presented with the following three different distribution options, with each offering unique advantages and disadvantages for your apps release: the Quest Store, App Lab, and SideQuest.

The Quest Store is the only native marketplace for games and apps.

App Lab games can be found in the Quest Store and on SideQuest.

SideQuest enables users to sideload your apps to their headsets.

Please note, there are other distribution avenues for your VR games, but this blog will focus solely on platforms accessible by Oculus Quest headsets. Other platforms will offer different distribution options with different submission processes.

The Quest Store

Of the three available, the Quest Store features the most rigid and demanding set of requirements that your app must meet in order to be published. Despite that, we recommend this route for developers publishing finished games they feel confident in; the Store offers unique and valuable advantages that will help your game find greater success such as excellent accessibility, little competition, and opportunities for organic discovery. These advantages, however, come at the cost of spending a significant amount of time optimizing your game to earn Oculus’ direct approval.

As far as app submissions are concerned, the Quest Store is characterized by its stringency. Apps distributed via the Store must comply with a whole host of requirements set forth by Oculus that ensure everything listed on the Store is of consistently high quality. Although several checks are merely recommendations, these checks span eleven different categories ranging from performance and functionality to security and accessibility.

The Quest Store’s quality-first mentality makes for a strict submission process, but it also reduces the potential competition your app will face should it earn Oculus’ approval. As of writing, there are only about 300 games listed on the Quest Store—less than half the amount available on App Lab, meaning Quest users browsing the Store are far more likely to organically find your app than users browsing App Lab’s library via SideQuest. The Quest Store is deeply integrated within Quest’s user-interface, so this channel also provides the easiest option for users looking to download and play games.

The Store reviews app submissions on a first-come, first-serve basis and has varying review times dependent upon the number of submissions at that time. A solid estimate is four to six weeks, but it’s worth noting there have been instances of developers waiting as long as three months or as little as two weeks.

App Lab

Games and apps published through App Lab are searchable in both the Quest Store and SideQuest. Think of this distribution channel as a happy medium between the Quest Store’s strict approval process and SideQuest’s comparatively lackadaisical approach to reviewing app submissions. That said, App Lab faces its own set of drawbacks; while this distribution avenue offers a more accessible means of entry, it also faces increased competition and minimal organic discovery.

App Lab titles are browsable via SideQuest, but these apps can only be found within the Quest Store by searching their full title exactly.

While App Lab submissions must abide by Oculus’ VRCs, several checks have been relegated from requirements to recommendations, making App Lab far more likely to approve your app than the Quest Store is. App Lab submissions are also reviewed on a first-come, first-serve basis and, although review times will vary, these windows are generally quicker than those of the Quest Store.

App Lab’s more straightforward approval process means that its library is growing far quicker than that of the Quest Store. With a current catalog comprising over 500 games and new ones being added every day, apps distributed via App Lab without any congruent marketing plan run the risk of sinking in this vast sea of options. Although SideQuest allows users to browse through the entire App Lab library freely, the Quest Store will only show these titles if you search their entire names precisely as listed. For example, searching Pavlov Shack through the Quest Store will not yield any results, but searching Pavlov Shack Beta will. Because so few people will naturally stumble across your App Lab game, you bear the responsibility of making sure your app gets seen.

SideQuest

SideQuest is essentially a giant, third-party store for Oculus Quest. It enables you to sideload games and apps onto your headset, provided you have access to a computer or mobile device. This distribution channel is the least accessible of the three as it also requires users to have their headsets in developer mode. However, unlike the Quest Store and App Lab, SideQuest submissions do not need to be approved by Oculus directly. Instead, these games and apps undergo a simpler, quicker administrative process that does not require your game to comply with any VRCs. SideQuest provides the quickest route for developers to share new builds with team members, playtesters, or friends.

Virtual Reality Checks to Keep in Mind

Getting an app approved by Oculus is easier said than done. Therefore, developers should be mindful of key VRCs as early as possible in order to ensure that major game design decisions are made with Oculus’ high standards in mind. After having undergone much of the optimization required to ready Loam Sandbox for official release, let’s briefly explore each of the eleven categories listed within the Virtual Reality Check guideline. This high-level overview of the requirements and recommendations will only apply to prospective Quest Store and App Lab games and apps.

1. Packaging

Packaging VRCs exist to ensure that every app available in the Quest Store and App Lab adheres to Oculus’ app packaging and formatting standards. There are six different required checks within this category regardless of whether you are submitting to App Lab or the Quest Store. One requirement, VRC.Quest.Packaging.5, requires that every submission be formatted as an APK file and smaller than 1 GB in size.

For a detailed look at the requirements that your app’s APK file must conform to, visit here.

2. Audio

The only Virtual Reality Check within this category is VRC.Quest.Audio.1, an optional recommendation for your app to support 3D audio spatialization features. Though not a requirement, incorporating audio spatialization within your game will mean that your app’s audio output will change as the user’s positioning does. We recommend that you consider implementing 3D audio within your games because doing so will add an immersive gameplay element, positively impacting player experiences.

3. Performance

Performance VRCs exist to ensure that the gameplay provided by titles available for Quest is of consistently high quality. This category comprises two App Lab requirements or three Quest Store requirements that concern your app’s refresh rate, runtime, and responsiveness.

Adhering to these checks can be difficult, thankfully there are several tools, such as the Developer Hub, that will enable you to accurately track the performance and overall functionality of your Quest apps. VRC.Quest.Performance.1 requires all App Lab and Quest Store listings to run at specified refresh rate minimums of 60 Hz and 72 Hz, respectively. You’ll want to keep this VRC in the back of your mind as you develop your game because its performance will benefit greatly from finding alternatives to CPU-intensive design choices.

We suggest using transparencies sparingly; whenever we made too many of Loam’s colors or textures see-through, we would experience drastic dips in frame rate. We also noticed that rigging critter skeletons within Unity would create a similar effect. To hear art optimization insights directly from our team, visit here.

We recommend sideloading OVR Metrics onto your headset and toggling on its persistent overlay setting. This will provide a real-time display of your app’s frame rate as you test its performance.

4. Functional

Oculus has created thirteen functional VRCs to ensure that all apps listed on App Lab and Quest Store run smoothly. Together, they prevent crashes, freezes, extended unresponsive states, and other game-breaking errors. Four of these checks are recommendations for App Lab submissions, while only one check is a recommendation for the Quest Store.

Developers seeking a spot in the Quest Store should keep VRC.Quest.Functional.1 in mind as they produce their games and apps. Because it requires you to be able to play through content for at least 45 minutes without any major errors, crashes, or freezing, this check could potentially impact your game’s design. VRC.Quest.Functional.1 is merely a recommendation for App Lab submissions, making that that ideal distribution avenue for early access games and demos not yet long enough for the Quest Store.

Visit here for a complete list of functional VRCs.

5. Security

Two security VRCs ensure that every Quest app respects and protects the privacy and integrity of customer data. Made up of one recommendation and a requirement, this is among the more straightforward categories. VRC.Quest.Security.2 is the only security check that developers must abide by; it requires your app to request no more than the minimum number of permissions needed for it properly function.

For more details, visit here.

6. Tracking

The tracking category of VRCs consists of one mandatory check for both Quest Store and App Lab submissions. To comply with VRC.Quest.Tracking.1, your app’s metadata must meet Oculus’ requirements for sitting, standing, and roomscale play modes. You can take in-depth look at this Virtual Reality Check here.

7. Input

Seven input VRCs relate to your app’s controls scheme and commands. This category is comprised of four and five requirements for App Lab and Quest Store, respectively; they function to maintain consistent interactivity from one Quest app to the next. For example, VRC.Quest.Input.4 makes sure that, when the user pulls up Quest’s Universal Menu, every Quest app continues to render in the background but hides the user’s hands.

If you plan on implementing hand tracking within your app, you’ll need to pay particular attention to this category. View all seven VRCs within the input category here.

8. Assets

Each of the eight VRCs included within this category is required of Quest Store submissions, but only half are needed for App Lab titles. These checks refer to the assets that will accompany your game within each distribution channel, ensuring that your app’s logos, screenshots, description, and trailer meet Oculus’ expectations. Though you may not be required to abide by every check if you’re looking to submit to App Lab, it’s important that you consider every check as you plan and create the marketing materials that will help your game stand out.

Clear cover art, for example, is only a requirement for the Quest Store, but this is an asset that will undoubtedly help your game stand out on App Lab too. Keep in mind that you will need multiple cover art images of different sizes and dimensions in order for your graphics to look their best in different areas of these distribution platforms. 

As you plan your game’s trailer, you should keep VRC.Quest.Asset.7 in mind—a required check limiting your trailer’s length to two minutes or less. While recording, take advantage of SideQuest’s ability to run ADB commands on your headset and alter the dimensions and bitrate settings of your Quest’s native screen-record feature. Also remember that, due to VRC.Quest.Asset.6, no asset accompanying your game can feature another VR platform’s headset or controllers.

9. Accessibility

Although none of these checks are requirements for either distribution channel, the accessibility category spans nine recommendations, making it among the largest sections in throughout the Quest Virtual Reality Check guideline. Failing to incorporate any of these suggestions won’t invalidate your submission, but we recommend developers implement these checks as well because they make sure that your game accommodates a variety of users.

Making adding subtitles, making text clear and legible, and incorporating color blindness options within your app’s settings are invaluable ways of making sure that it is accessible to the broadest audience possible. For ever suggestion within this category, visit here.

10. Streaming

Streaming VRCs exist to guarantee all Quest apps are capable of providing smooth streaming experiences. VRC.Quest.Streaming.2 is the only required streaming check, making streaming among the simpler categories to comply with. You’ll just need to make sure that your app can only stream VR content to local PCs that the user has physical access to.

11. Privacy

Five required privacy checks ensure that all games comply with Oculus’ Privacy Policy requirements. Your app’s Privacy Policy must clearly explain what data is being collected, what it is being used for, and how users may request that their stored data be deleted. All five checks included within this category are requirements for titles on either the Quest Store or App Lab.

Click here for more information regarding Oculus’ Privacy Policy Requirements.

Have you submitted an app through the Quest Store, App Lab, or SideQuest? Are you planning to do so? Let us know what your submission process was like by leaving a comment below or tagging us on Twitter: @AmebousLabs

]]>
GAME ART DESIGN FOR VR: FORM AND FUNCTION https://amebouslabs.com/game-art-design-for-vr-form-and-function/ Tue, 26 Oct 2021 20:33:56 +0000 https://amebouslabs.com/?p=3851

GAME ART DESIGN FOR VR: FORM AND FUNCTION

Annie Eaton

Written by Nick Foster

October 26, 2021

Designing art for virtual reality games is to wrestle form with function. All things naturally have a form or shape–an outward appearance that differentiates one item from the next. Similarly, all art has a purpose or function, whether expressive, utilitarian, or persuasive in nature. Louis Sullivan, an influential architect from the late 19th century, believed that “form follows function.” Of course, he did not create his philosophy with VR game design in mind, but his ideas continue to be altered, debated, and applied to all forms of art today.

As we near the release of Loam Sandbox, we can begin to reflect upon the lessons learned while designing the VR gardening game’s art and consider how Sullivan’s guiding tenet influenced the choices, sacrifices, and concessions we had to make to get Loam looking and functioning how we wanted it to.

Sullivan pioneered the American skyscraper; he believed that each skyscraper’s external “shell” should reflect its internal functions. 

Loam began with a purpose; we wanted to make a relaxing gardening simulation game in virtual reality. With that in mind, we started researching environments that would take advantage of the immersion provided by Oculus Quest headsets without inhibiting the game’s overall goal.

From the very start, Loam’s form followed its function. We ultimately decided that Loam’s setting would be a mountain valley region because the setting facilitates realistic gardening while providing a vast landscape that shines when showcased in virtual reality. Even when developing non-VR games, it is essential to consider how your game’s setting interrelates with its core gameplay.

Now, with a basic understanding of Loam’s setting, we continued to research mountain valley regions and their associated cultures. Knowing that we wanted this game to be a relaxing respite for players, we used this research to hone in on the tone and feel of the game.

Hygge is a popular theme across Scandinavian cultures but is especially prominent in Denmark and Norway. Because this idea of comfort and the feelings of well-being that accompany it aligned so heavily with the function of Loam, it inevitably had a significant impact on the visual form of everything we see in Loam, from its architecture to the game’s textures and assets.

 

As we grappled Loam’s comforting visuals with its gardening gameplay, we also had to consider the game’s functionality as a whole in terms of its ability to run smoothly on both Oculus Quest and Quest 2 headsets. Before either the Oculus Quest Store or App Lab can distribute a game or app, it must meet the stringent Virtual Reality Check (VRC) guidelines set forth by Oculus.

These VRCs are yet another example of the role Loam’s functionality played in determining its visual form. These guidelines exist to ensure that all games on the platform are of a consistently high standard; unfortunately, meeting these guidelines often requires the alteration or removal of certain elements and assets.

Your game’s performance is a common reason you might retroactively modify your game’s form to align with its function, but another critical aspect to consider is the player’s experience, especially when developing virtual reality games.

Playtesting is an invaluable opportunity to ensure your game is not off-putting to players. Feedback from players will help you better align your game’s form with its function. By playing your game vicariously through a playtester’s fresh eyes, you’ll discover aspects of your game that should be improved, removed, or implemented.

Join our discord to keep up to date with every change we make to the world of Loam leading up to the release of Loam Sandbox.

]]>
KICKSTART YOUR CAREER IN GAME DEVELOPMENT https://amebouslabs.com/kickstart-your-career-in-game-development/ Mon, 20 Sep 2021 22:03:12 +0000 https://amebouslabs.com/?p=3685

KICKSTART YOUR CAREER IN GAME DEVELOPMENT

Annie Eaton

Written by Nick Foster

September 20, 2021

Let’s face it, we all want to make games. Game development offers a uniquely rewarding career that allows you to celebrate your work in unusual ways. In what other industry are you afforded the luxury of being able to literally play with your work? Luckily, every great game needs a team behind it—people who can design gameplay, implement sounds, write code, create assets, and playtest each build. Because game production blends technical abilities and logic with artistic expression and creativity, there’s a role for everyone regardless of your skillset or interests. Take our team, for example:

With so many accessible, often free online resources at your disposal, there’s never been a better time to enter the gaming industry. However, just like any other significant undertaking, figuring out where to begin is the hardest part. Whether you want to make game development your new career or your next hobby, our team of developers and artists will share within this article their go-to tools and resources, plus their suggestions for professionals looking to break into the industry.

With so many resources at your disposal, there’s never been a better time to enter the gaming industry.

Our team’s best advice for an aspiring game artist or developer would be to start small and think big. If you’re interested in forging a career in game development, then you should be prepared to spend time developing your abilities and familiarizing yourself with the software and languages you anticipate using the most. With that said, the best place for you to start your journey might actually be away from the computer.

You can’t start designing games until you start thinking like a game designer! Creating your own board game is an incredibly effective introduction to the world of game production. When you look at the steps involved, producing a board game is surprisingly similar to video game production. Regardless of their medium, every game needs a working set of rules designed with players in mind, then they eventually need playtesters to highlight potential improvements for the next iteration or build.

At the moment, many aspiring 3D artists and developers are denied the hardware needed to develop satisfying video games simply because it is too expensive to acquire. This is another reason board games are a great place to start, but if you’re truly ready to take the next step, we suggest taking advantage of the local opportunities unique to your area. By participating in meetup groups and organizations like the International Game Developer’s Association (IDGA), you may find regular access to the expensive hardware you’re seeking.

To craft the magical world of Loam, our team uses the game engine Unity, which you can download and start learning for free. As you familiarize yourself with Unity (or your engine of choice) and start producing your game, you will inevitably encounter obstacles. Bringing questions to your local meetup group is a great way to gain new insights, but this is not always possible. Thankfully, there are a plethora of free and accessible resources online!

Meeting others with similar interests, skills, and goals is perhaps the most exciting aspect of pursuing any passion. While attending college facilitates limitless networking opportunities, it is not the only path for aspiring developers or artists. Above all else (including degrees), you should value your portfolio, your previous work experience, and your connections in the industry. With so many possible ways to enter the industry, here is how our team managed to do so:

Aspiring game artists and developers should be prepared to spend lots of time refining their trade. Again, start small and think big; with consistent practice, you’ll start to master your software, understand programming languages, improve your artistic vision, and perfect your creative abilities.

]]>