This is the first of a series of blogs about using Swift, Apple's new programming language that was annouced at WWDC 2014. They are not in any particular order but are about new and interesting features and differences with Objective-C. I don't intend to explain syntax or specific frameworks as there are plenty of resources available covering those already.
When developing iOS apps a number of things are required for producing a quality product. These include the ability to run unit tests on the code and to test the user interface automatically. We can set up Continuous Integration (CI) to complete these tests for us. If all tests are passed then we automatically upload to Hockeyapp for our testers to download. In this blogpost I'll take you through the process of setting up a CI pipeline for iOS.
We built a very simple app called ‘Tasky’ which pretty much does as little as possible but enough to be able to build a complete workflow. We wanted an app with a bit of navigation and Core Data to give us enough to test.
I’m going to share some of the thinking and code behind the Picle iPhone app starting with the camera functionality. This will involve showing some code snippets and describing classes and frameworks found in the Apple iOS SDK.
At the heart of the the Picle app is the ability to use the iPhone camera and microphone in quick succession. The user experience of this functionality is critical to feel of the app so plenty of work has been done both before and after the initial release. Indeed the next release will have this completely reworked.
Anyone who has used iOS will be familiar with the way Apple uses animation in their apps. It's one of the most delightful features of the platform and users loved it, even before there was an App Store. If you use apps that don't employ judicious animation, you get a sense of something missing, of an undefinable lack of quality.