In the world of tech, transitioning from the role of an engineer to an engineering manager is often viewed as a natural progression. Many engineers, after having honed their technical skills, look forward to the opportunity of expanding their sphere of influence by stepping into a managerial role. However, this transition is not as straightforward as it may seem at first glance. It presents a new set of challenges that are fundamentally different from those encountered in a purely technical role.
Understanding the Role Shift: From Problem-Solver to Enabler
One of the most significant challenges in transitioning from engineer to engineering manager is accepting and understanding the fundamental shift in roles. As an engineer, your primary role is to solve technical problems - whether it’s writing code, debugging issues, or designing software architecture. Your success is often measured by the quality of the solutions you deliver and the technical challenges you overcome.
However, as an engineering manager, your role involves less hands-on problem-solving and more enabling your team to solve problems. You become a facilitator, mentor, and guide, helping to clear roadblocks and provide resources for your team. Your success is now measured less by your personal technical accomplishments and more by the achievements of your team.
This shift from individual contributor to team leader can be challenging and even discomforting for many new managers. It requires a change in mindset, where you need to redefine what productivity and success mean to you. You’ll need to find fulfillment in the success of others, which can be a difficult transition for someone used to personal technical achievements.
Mastering the Art of Time Management and Prioritization
As an engineer, you are often given well-defined tasks and deadlines. Your work is scoped, and you can focus your efforts on technical problem-solving. However, as a manager, your responsibilities become varied and often less defined. You’ll be juggling multiple tasks, from strategic planning and goal-setting to recruitment, team-building, and administrative duties.
Consequently, effective time management and prioritization become essential skills. You’ll need to learn to balance urgent issues with important long-term goals, delegate tasks, and sometimes, even say ‘no’ to avoid overcommitting yourself or your team.
Navigating the Human Element: People Management
Perhaps one of the biggest challenges for many new engineering managers is people management. While your previous role primarily involved interacting with code, your new role will involve interacting with people. Understanding and managing human dynamics is much more complex and unpredictable than solving technical problems.
You’ll need to cultivate a diverse set of skills, including effective communication, conflict resolution, and motivational abilities. You’ll have to navigate team dynamics, manage conflicting personalities, understand individual motivations, and foster a positive, productive work environment. You’ll also need to provide feedback, mentorship, and career development for your team members - tasks that require empathy, patience, and understanding.
The Challenge of Delayed Feedback
In engineering, feedback is often immediate and clear-cut. You write code, run it, and see the results. If something breaks, you fix it. Your impact is visible and quantifiable. However, as a manager, feedback is often delayed and less tangible.
The impact of decisions you make or initiatives you implement may not be evident for weeks or even months. Measuring success becomes trickier, as it’s often tied to team performance, employee satisfaction, and long-term project outcomes. This delay in feedback can be disconcerting and requires patience, long-term thinking, and an ability to see the big picture.
Growing into a Visionary Leader
As an engineer, you are typically tasked with implementing visions and plans created by others. However, as an engineering manager, you are expected to set the vision and chart the course. This involves strategic thinking, decision-making under uncertainty, and the ability to inspire and motivate your team to align with your vision. It requires you to shift from a detail-oriented focus to a broader perspective, which can be challenging for many who are used to the concrete world of coding.
The Isolated Landscape of Management
Being a manager can sometimes feel isolating. You’re often privy to sensitive information that you can’t share with your team. There may be company decisions you have to uphold even if you don’t personally agree with them. You may need to mediate conflicts, deliver difficult feedback, or make tough decisions that not everyone will like. This can create a sense of being ‘alone in the middle’ - between your team and upper management - which many new managers are unprepared for.
The Conquest of the Challenges: Strategies for Success
While these challenges might seem daunting, they can be successfully managed with the right strategies.
Embrace the Change: Recognize that your role has fundamentally changed. Embrace your new identity as a facilitator and leader, and take pride in the success of your team.
Develop New Skills: Invest time in developing essential managerial skills, such as effective communication, conflict resolution, strategic thinking, and time management. These are just as important as your technical skills in your new role.
Seek Mentorship: Find mentors who can guide you through this transition. They can provide invaluable advice, insights, and moral support.
Prioritize Self-Care: The role of a manager can be stressful. Make sure to take care of your mental health. Practice mindfulness, maintain a healthy work-life balance, and seek professional help if needed.
Embrace Continuous Learning: Accept that you’ll make mistakes, but view them as learning opportunities. Strive to learn and grow continuously.
Final Thoughts
Transitioning from engineer to engineering manager is a significant career shift, filled with challenges but also opportunities for personal and professional growth. It’s a journey that requires not just technical expertise but also a broad range of soft skills.
However, it’s important to remember that management is not the only path for career progression. Many companies now offer ‘individual contributor’ tracks, allowing engineers to take on more complex problems and responsibilities without moving into management.
Whether you decide to pursue management or remain an individual contributor, the key is to understand your strengths, passions, and career aspirations, and align them with your chosen path. Remember, a successful career is not just about titles or roles, but about continuous learning, growth, and making a positive impact in your organization.
In the end, whether you’re writing code or leading a team, the most important thing is to find joy and fulfillment in what you do. This will not only lead to personal satisfaction but also to a successful and rewarding career.
In this series of posts we are going to be covering the SOLID principles of software development. These are a set of principles / guidelines, that when followed when developing a software system, make it more likely that the system will be easier to extend and maintain over time. Let’s take a look at the problems that they seek to solve:
Fragility: A change may break unexpected parts, it is very difficult to detect if you don’t have a good test coverage
Immobility: A component is difficult to reuse in another project or in multiple places of the same project because it has too many coupled dependencies
Rigidity: A change requires a lot of effort because it affects several parts of the project
So what are the SOLID principles?
Single Responsibility Principle - A class should have only a single responsibility / have only one reason to change
Open-Closed Principle - Software should be open for extension but closed for modification
Liskov Substitution Principle - Objects in a program should be replaceable with instances of their sub types without altering the correctness of the program
Interface Segregation Principle - Many client-specific interfaces are better than one general-purpose interface
Dependency Inversion Principle - High level modules should not depend on low level modules. Both should depend on abstractions
In this article we will focus on the Dependency Inversion Principle.
What does it mean?
This principle has 2 main components described below:
High-level modules should not import anything from low-level modules. Both should depend on abstractions (e.g., interfaces)
Abstractions should not depend on details. Details (concrete implementations) should depend on abstractions
I’ve interviewed many iOS developers over the years and I struggle to recall a single person who has actually got this part of the SOLID principles 100% right. I think a large part of this comes from the fact many use simple architectures such as MVVM that don’t break applications down into smaller layers and components/modules. This isn’t a criticism, not every app needs a VIPER/Clean architecture approach with multiple layers.
Most iOS developers I speak to come to the conclusion that this principle just means using protocols instead of concrete classes and injecting dependencies to help with testing/mocking. While this principle does rely on this to work it is not the primary purpose of the principle and exposes an issue once you start to depend on abstractions used across multiple layers / modules.
Setup
Lets setup a simple example in Xcode where we have two separate modules that depend on each other in order to provide some functionality.
If we create a simple Swift UI project using Xcode, then using File -> New -> Package create 2 new swift packages and add them to the project. One called LibraryA and one called LibraryB. Be sure to select your project in the ‘Add to’ drop down when naming your libraries. You should have something that looks like below Xcode.
Let’s start by adding a protocol and some custom data structure that we are going to be using across the 2 libraries we are working. Add a file called Protocol.swift in LibraryA and add the following code below:
This is just a simple class that has the MyProtocol protocol has a dependency. Simple enough so far!
Now let’s create a class in LibraryB that implements the protocol that is being used in our class in LibraryA. Create a class in LibraryB called ImplementationB:
We have our class in LibraryB that is implementing the protocol we previously created in LibraryA. For this reason you will notice that we have to import LibraryA in this class as well. There is one additional step we need to do before this will compile correctly, we need to define our dependency in our package file. Let’s open the LibraryB package file and edit it to add the dependency between the 2 packages:
importPackageDescriptionletpackage=Package(name:"LibraryB",products:[.library(name:"LibraryB",targets:["LibraryB"]),],// Assign LibraryA as dependencydependencies:[.package(path:"../LibraryA")],targets:[.target(name:"LibraryB",// Add dependency to targetdependencies:["LibraryA"]),.testTarget(name:"LibraryBTests",dependencies:["LibraryB"]),])
We won’t go into all the options you need in the swift package file but hopefully by reading this you can see we have defined a dependency and added it to our LibraryB target. If you attempt to build the project it should compile successfully.
Now let’s look at the structure of these 2 libraries and their relationship to each other.
As you can see, we have LibraryA, this contains the protocols and LibraryB that implements the protocols. In order for LibraryB to implement the protocols it needs a dependency to LibraryA in order to work.
Problem 1
Now what happens if we want to use the protocols in another Library? Let’s call this LibraryC? At moment the protocols are contained in LibraryA where they are being used and implemented by Library B.
We can’t use the protocols in another library without adding LibraryA which may contain code and other assets we don’t want.
We could copy the protocols to LibraryC, however if you needed LibraryA and LibraryC in the same project you would get class clashes.
We would need to edit LibraryB to add another dependency in this case. What happens if we are not the owners of LibraryB? How would we even do this?
One solution we can try is moving the protocols from LibraryA to LibraryB. This reverses the dependencies and helps to solve the problems highlighted above. Let’s go ahead and do this now.
Copy the Protocols.swift file we created from LibraryA to LibraryB
Remove the import of LibraryA from the implementationB.swift
Update the LibraryB package file to remove the LibraryA dependency
importPackageDescriptionletpackage=Package(name:"LibraryB",products:[.library(name:"LibraryB",targets:["LibraryB"]),],// Assign LibraryA as dependencydependencies:[],targets:[.target(name:"LibraryB",// Add dependency to targetdependencies:[]),.testTarget(name:"LibraryBTests",dependencies:["LibraryB"]),])
Update the LibraryA package file to add the LibraryB dependency
importPackageDescriptionletpackage=Package(name:"LibraryA",products:[// Products define the executables and libraries a package produces, and make them visible to other packages..library(name:"LibraryA",targets:["LibraryA"]),],dependencies:[// Dependencies declare other packages that this package depends on..package(path:"../LibraryB")],targets:[// Targets are the basic building blocks of a package. A target can define a module or a test suite.// Targets can depend on other targets in this package, and on products in packages this package depends on..target(name:"LibraryA",dependencies:["LibraryB"]),.testTarget(name:"LibraryATests",dependencies:["LibraryA"]),])
Now let’s see a diagram of what we have done. Now if we look at our 2 libraries at a high level the dependencies look like this.
So now, our LibraryA has a dependency on LibraryB. LibraryA is using the protocols that are defined AND implemented in LibraryB! Problem solved… right… Not entirely.
If we review the problems we discussed earlier, if LibraryC wanted to make use of the protocols and implementation in LibraryB that is now possible as LibraryB has no knowledge of LibraryA or LibraryC. However this creates a new problem…
Problem 2
What if LibraryA and LibraryC want to use different implementations of the protocols from each other? What if we introduced LibraryD that wanted to implement the protocols and be used by another library such as LibraryC? In order to facilitate this we would need to create a dependency between LibraryD and LibraryB. What problems does this create?
We are introducing a dependency to a library we don’t need, in order to implement the protocols within it. In our example there isn’t much in LibraryB but imagine LibraryB had lots of other code and its own dependencies? Now we are including all of those in our project in order to have access to the protocols.
Solution
This is where Dependency Inversion comes in. What we need to do is create a separate library for the protocols and any data that passes through them. Then, we make the dependencies between our different libraries to that one, thus removing the dependencies between our different layers. Let’s do that now.
Create a new package and add it to your project, naming it Protocols. Now move the Protocols.swift file that we created earlier into the Protocols package. Your Xcode project file explorer should look like below:
Now lets edit the dependencies of our packages so that both LibraryA and LibraryB depend on protocols. Your package files for LibraryA and B should now look like the below:
importPackageDescriptionletpackage=Package(name:"LibraryA",products:[// Products define the executables and libraries a package produces, and make them visible to other packages..library(name:"LibraryA",targets:["LibraryA"]),],dependencies:[// Dependencies declare other packages that this package depends on..package(path:"../Protocols")],targets:[// Targets are the basic building blocks of a package. A target can define a module or a test suite.// Targets can depend on other targets in this package, and on products in packages this package depends on..target(name:"LibraryA",dependencies:["Protocols"]),.testTarget(name:"LibraryATests",dependencies:["LibraryA"]),])
importPackageDescriptionletpackage=Package(name:"LibraryB",products:[// Products define the executables and libraries a package produces, and make them visible to other packages..library(name:"LibraryB",targets:["LibraryB"]),],dependencies:[// Dependencies declare other packages that this package depends on..package(path:"../Protocols")],targets:[// Targets are the basic building blocks of a package. A target can define a module or a test suite.// Targets can depend on other targets in this package, and on products in packages this package depends on..target(name:"LibraryB",dependencies:["Protocols"]),.testTarget(name:"LibraryBTests",dependencies:["LibraryB"]),])
Now update the imports of your implementation files so that they import Protocols instead. Your implementation files should look like the below:
Let’s have a look at what we have done here. We have moved the dependencies between the layers into a separate package and are now pointing both sides at that rather than one way or the other. Let’s update our diagram to show how this helps us.
Now if we want to use LibraryA, B, C, or D it does not matter in our dependency graph. They all point to the protocols and data and can be used interchangeably without the need to modify the libraries, so they depend on each other. We also avoid importing any unnecessary classes that we don’t need in order to satisfy the dependencies.
This is what dependency inversion is. It’s separating protocols and data dependencies from the dependencies themselves and putting them into their own package. This way you completely separate the layers from each other, and they can be used together without any knowledge of each other. Awesome!
In this series of posts we are going to be covering the SOLID principles of software development. These are a set of principles / guidelines, that when followed when developing a software system, make it more likely that the system will be easier to extend and maintain over time. Let’s take a look at the problems that they seek to solve:
Fragility: A change may break unexpected parts, it is very difficult to detect if you don’t have a good test coverage
Immobility: A component is difficult to reuse in another project or in multiple places of the same project because it has too many coupled dependencies
Rigidity: A change requires a lot of effort because it affects several parts of the project
So what are the SOLID principles?
Single Responsibility Principle - A class should have only a single responsibility / have only one reason to change
Open-Closed Principle - Software should be open for extension but closed for modification
Liskov Substitution Principle - Objects in a program should be replaceable with instances of their sub types without altering the correctness of the program
Interface Segregation Principle - Many client-specific interfaces are better than one general-purpose interface
Dependency Inversion Principle - High level modules should not depend on low level modules. Both should depend on abstractions
In this article we will focus on the Interface Segregation Principle.
What does it mean?
The summary of the principle is as follows:
Many client-specific interfaces are better than one general-purpose interface
In Swift, we use Protocols rather than interfaces in languages such as Java so from here on out we will refer to interfaces as Protocols.
Now the purpose of this rule is quite straight forward in comparison to some of the other rules in the SOLID principles. What it means is its better to create smaller Protocols than to create one large one with lots of methods defined.
What’s the problem
So why does having one large protocol cause a problem? Let’s examine one of the classic Cocoa Touch protocols to see why this is an issue.
I am sure many of you have implemented this protocol at some point in your past ;) I have modified the source slightly to make it easier to read and get the point across So why are we looking at this?
Only 2 methods you have to implement on the first 2.
The rest of the methods are optional and you can implement whichever ones you feel you want to use.
Now this protocol has its roots in Objective C helps it masks the problem somewhat. In Objective C as you can see in the code above its possible to mark certain functions as optional. This means you can implement them if you want to but don’t have to, this allows this protocol declaration to contain too many methods without causing problems for the implementing class.
In Swift, it is not possible to mark functions as optional, all functions need to be implemented. Let’s update the above protocol to be more Swifty and see what problems that might cause us.
Now that we have converted our protocol to be more swifty, what problem will this cause when we attempt to make a class conform to this protocol? Let’s have a look at an example.
Our class above now has to implement every single protocol method. Even if we don’t intend to use it. In the objective c implementation of the protocol we have the option of implementing only the ones we need whereas now we must implement every single method. Imagine all the view controllers in the world that would be full of empty methods in order to conform to this protocol!
This protocol breaks the interface segregation principle.
A better solution
To improve the solution we could break the one big interface down into smaller protocols. That way we could conform to only the protocols we were interested in implementing for our functionality. This may looks something like:
UITableViewDataSource - For the 2 compulsory methods we are familier with
UITableViewSectionsDatasource - For methods relating to multi section methods
UITableViewSectionTitles - For methods relating to headers and footers in sections
UITableViewEditable - For methods relating to editing and moving cells
This way we could conform to select methods we want, rather than one big interface where we may only want a small subset of the methods.
A good example
A good example of interface segregation in the iOS SDK is Codable. The definition of Codable is as below:
typealiasCodable=Decodable&Encodable
Basically Codable is just the combination of 2 other protocols: Decodable and Encodable. This is a great example of how to do the interface segregation. If you are building say a JSON parse struct, you may wish to only conform to Decodable so you can decode the JSON. If in the future you wanted to serialize the struct for something like say data storage you can conform to Encoding later if needed.
Summary
The interface segregation principle is the easiest of the principles to understand in my opinion. In basic terms it means don’t create a big protocol with lots of methods that aren’t always required to be implemented depending on the implementation requirements.
Instead, separate the protocol into smaller protocols with only the methods required for a single piece of functionality to work. Not only does this avoids having lots of redundant methods it also helps to facilitate the single responsibility principle by allowing functionality to be broken down into different classes. For example, you could have different classes to handle different activities rather than one big class with all functionality in.
In this series of posts we are going to be covering the SOLID principles of software development. These are a set of principles / guidelines, that when followed when developing a software system, make it more likely that the system will be easier to extend and maintain over time. Let’s take a look at the problems that they seek to solve:
Fragility: A change may break unexpected parts, it is very difficult to detect if you don’t have a good test coverage
Immobility: A component is difficult to reuse in another project or in multiple places of the same project because it has too many coupled dependencies
Rigidity: A change requires a lot of effort because it affects several parts of the project
So what are the SOLID principles?
Single Responsibility Principle - A class should have only a single responsibility / have only one reason to change
Open-Closed Principle - Software should be open for extension but closed for modification
Liskov Substitution Principle - Objects in a program should be replaceable with instances of their sub types without altering the correctness of the program
Interface Segregation Principle - Many client-specific interfaces are better than one general-purpose interface
Dependency Inversion Principle - High level modules should not depend on low level modules. Both should depend on abstractions
In this article we will focus on the Liskov Substitution Principle.
What does it mean?
So the Liskov Substitution Principle states:
Derived classes must be substitutable for their base classes
What exactly does this mean? In a basic sense it for example if you have a function that accepts a type of class which is a parent of other classes, any class that subclasses the parent class should be able to be passed in without it breaking the program.
See a summary of the main points of the principle below:
Contra variance of method parameter types in the sub type.
Covariance of method return types in the sub type.
New exceptions cannot be thrown by the methods in the sub type, except if they are sub types of exceptions thrown by the methods of the super type.
Don’t implement stricter validation rules on input parameters than those implemented by the parent class.
Apply at the least the same rules to all output parameters as applied by the parent class.
Let’s take a look at what these different rules mean for subclasses.
The Parent Class
First of all, let’s define our parent class or base class that contains some functionality. Let’s use a vehicle class as an example, this vehicle has a throttle which can be set at any value between 0 and 100.
First of all we define a custom error to throw if the throttle is not within bounds
Here we define our vehicle class that has a throttle variable to store the value being set
We have a function to set the throttle value, there is a guard statement to check whether the value being set is in the appropriate range. If it is not, we throw an error, if it is we set the value
Validation rules on input parameters
Now let’s create a subclass that breaks the principle. We will make a lorry class that inherits from the super class but adds its own restrictions to the throttle function, only allowing the throttle to be set between 0 and 60 for example.
So what is happening here? We have subclassed the Vehicle class and overriden the setThrottle method. Now what we have done here is we have added a guard statement to check if the throttle is between 0 and 60. We throw an error saying out of bounds if outside of that, if it is within bounds we call the super class method.
Why is this a problem? Well imagine we are building a system / class that interacts with the Vehicle class. Now based on the Vehicle class you would expect to be able to set the throttle to anything between 0 and a 100. However now, if someone chooses to pass a Lorry subclass to your system / class, you will not be able to set the throttle above 60. Depending on how this other class or system is built this may have unintended side effects as you can’t set the values that you are expecting without getting an error.
This example breaks the rule:
Don’t implement stricter validation rules on input parameters than those implemented by the parent class.
Errors in the liskov principle
Let’s modify our example to see how we could break the principle by throwing different errors. Let’s modify the Lorry subclass:
We are calling our function with a value considered out of bounds
We catch the outOfBounds exception and print a system shutdown message
We have a generic catch for other errors where we show a generic error message
Now if we run this code we see the below message in the console as expected:
System shutdown
So what happens if we replace our Lorry subclass with its new error and put it in place of the Vehicle super class? If we change line one to read:
letvehicle:Vehicle=Lorry()
If we run the code above we will now see a different error:
Show generic error
The error handling code is not aware of subclass specific errors so is no longer able to handle them accordingly. Imagine a mission critical system that needs to shut down if an out of bounds happens, in this case the error would be missed as it would require the error handling class to have knowledge of all possible sub types in order to handle all the errors appropriate. Defeating the point of using the super class and thus breaking the principle:
New exceptions cannot be thrown by the methods in the sub type, except if they are sub types of exceptions thrown by the methods of the super type.
Contra variance and Covariance of parameters and return types
In the list of rules you may recall seeing two items talking about contra variance and covariance of parameters and return types. What does that mean exactly?
Contra variance of method parameter types in the sub type.
Covariance of method return types in the sub type.
Contra variance of method parameter types in the sub type
Contra variance means that we can use change the type of method parameter to a super class of the type but not a subclass. This rules works basically in combination with the rule below:
Don’t implement stricter validation rules on input parameters than those implemented by the parent class.
What it means is, we can use a super class of a parameter, thus ‘weakening’ the restrictions of the method, but not a subclass of that type which would tighten the restrictions of the method.
Covariance of method return types in the sub type
Covariance means that the type being used can be a sub type of the class provided by the super class function return type. Similarly, this works in the same way as the 5th rule:
Apply at the least the same rules to all output parameters as applied by the parent class.
Now both of these rules aren’t possible to be broken as part of Swift. It’s not possible to overload functions providing alternative type specifications, at least while still referring to the super class type. We can override methods and provide different parameter types and return types but this requires the calling class to know the type of the subclass. When referring to the super class, the super class implementation is always called regardless of subclass functions with different params.
In this series of posts we are going to be covering the SOLID principles of software development. These are a set of principles / guidelines, that when followed when developing a software system, make it more likely that the system will be easier to extend and maintain over time. Let’s take a look at the problems that they seek to solve:
Fragility: A change may break unexpected parts, it is very difficult to detect if you don’t have a good test coverage
Immobility: A component is difficult to reuse in another project or in multiple places of the same project because it has too many coupled dependencies
Rigidity: A change requires a lot of effort because it affects several parts of the project
So what are the SOLID principles?
Single Responsibility Principle - A class should have only a single responsibility / have only one reason to change
Open-Closed Principle - Software should be open for extension but closed for modification
Liskov Substitution Principle - Objects in a program should be replaceable with instances of their sub types without altering the correctness of the program
Interface Segregation Principle - Many client-specific interfaces are better than one general-purpose interface
Dependency Inversion Principle - High level modules should not depend on low level modules. Both should depend on abstractions
In this article we will focus on the Open-Closed Principle.
What does it mean?
So the open-closed principle states:
Software should be open for extension but closed for modification
What exactly does this mean? I think out of all the principles this is the hardest to understand. Mostly due to the fact the explanation leaves far too much open to interpretation. A simple Google search will offer up several examples of how this principle works. In this article I will present my take on the principle and how to build software that will comply with it.
Let’s focus on a type that by design, violates the open closed principle.
Enums
Enums are a very powerful tool in Swift. They are first class types and such can have associated values and conform to protocols for example. However when used at the boundary of a particular system or module they can present a particular problem.
Let’s imagine an analytics system where we can log events. This is a design pattern I’ve seen in many places:
// 1enumAnalyticsEvent{casenewsListcasenewsDetail(id:Int)}// 2classAnalyticsController{funcsendEvent(_event:AnalyticsEvent){lettitle=event.titleletparams=event.params// Send data to analytics network}}// 3extensionAnalyticsEvent{vartitle:String{switchself{case.newsList:return"News List"case.newsDetail:return"News detail"}}varparams:[String:String]{switchself{case.newsList:return[:]case.newsDetail(letid):return["id":"\(id)"]}}}
Let’s look at what we have here.
The first thing that is defined is an enum that houses all the different analytics events that are available. Some with associated values.
Next we have our analytics controller, this takes an event as a parameter, takes information from the event and would then send that on to our analytics system.
Here we have extended the AnalyticsEvent enum to add 2 variables, one for title and one for params that contain a switch for each of our events.
On the surface or at first look this might appear an ok solution. We have hidden our implementation of the analytics network inside our AnalyticsController and setup a defined set of events that we can support.
The Problem
Now lets look at the problems that this approach causes.
What happens if we need to add new events to our analytics system?
What if our analytics system was part of a separate package or module?
What happens when we have a lot of events?
So first of all, every time we need to add / update or remove any of the events in our analytics system we need to modify the enum. We can’t just implement new events and have them be compatible with the system. Also if the number of events becomes very large then the code will grow large in size. Making it hard to read, maintain and a bit of a mess. Also the enum now has multiple responsibilities, as it covers many events breaking the single responsibility principle.
The second issue which is probably the more important one, is let’s say we are breaking our app down in to separate packages. This Analytics Controller and Event would be in a separate package, what if we wanted to re-use it across different projects? Both of these scenarios become very difficult because we are using an enum that would need to be updated to accommodate events for different apps. The package would need constantly updating as new events were added.
The Solution
So we have identified some issues with the above implementation, how can we change it to make solve these issues we have identified? Look at the new example:
// 1structNewsListEvent:AnalyticsEvent{vartitle:String{return"News List"}varparams:[String:String]{return[:]}}structNewsDetailEvent:AnalyticsEvent{letid:Intvartitle:String{return"News detail"}varparams:[String:String]{return["id":"\(id)"]}}// 2protocolAnalyticsEvent{vartitle:String{get}varparams:[String:String]{get}}classAnalyticsController{funcsendEvent(_event:AnalyticsEvent){lettitle=event.titleletparams=event.params// Send data to analytics network}}
Let’s look at how we have changed the example:
First of all we have now removed the enum. Using an enum as a huge list of possible options is considered a code smell. Especially when it involves something that may change often. If you have a finite number of states that is unlikely to change, that is more suited to an enum than a list of analytics events. We have refactored those enum cases into 2 separate classes now.
We have switched the enum for a protocol that exposes the necessary items required by our analytics controller (we could have potentially done this in the previous example however we would still have the enum).
So what advantages does this have over the previous implementation?
With the events now being in separate classes we are now following the single responsibility principle, each event has its own class that can be updated whenever they need to be.
Now that we are using a protocol and not an enum, we are now able to add new events to our app without ever having to touch the analytics system. Simply create a new class and make it conform to AnalyticsEvent, and we can use it with the analytics controller.
Further to that we could have our analytics system in a separate reusable package, then our client apps could define their own set of events to use with the system.
Our analytics code is now open for extension, but does not need to be modified to support new events. Unlike our enum example.
In this series of posts we are going to be covering the SOLID principles of software development. These are a set of principles / guidelines, that when followed when developing a software system, make it more likely that the system will be easier to extend and maintain over time. Let’s take a look at the problems that they seek to solve:
Fragility: A change may break unexpected parts, it is very difficult to detect if you don’t have a good test coverage
Immobility: A component is difficult to reuse in another project or in multiple places of the same project because it has too many coupled dependencies
Rigidity: A change requires a lot of effort because it affects several parts of the project
So what are the SOLID principles?
Single Responsibility Principle - A class should have only a single responsibility / have only one reason to change
Open-Closed Principle - Software should be open for extension but closed for modification
Liskov Substitution Principle - Objects in a program should be replaceable with instances of their sub types without altering the correctness of the program
Interface Segregation Principle - Many client-specific interfaces are better than one general-purpose interface
Dependency Inversion Principle - High level modules should not depend on low level modules. Both should depend on abstractions
In this article we will focus on the Single Responsibility Principle.
Problem
The first principle in the list is the Single Responsibility Principle. This principle is defined as follows:
A class should have only one reason to change
This means a class should be responsible for only one task, not multiple. Let’s take a look at an example and how we can refactor it using the principle.
This looks like a fairly simple news datasource / service that is fetching some news items from the web. However if we take a closer look we will see it’s responsible for more than one task.
Creating the URLRequest that is used to fetch the news articles
Fetching the data using a URLSession
Parsing the data
Already that is 3 different responsibilities this class has. They may seem fairly straight forward in this example but imagine how this could get out of hand quickly in a larger codebase. Let’s cover some of the scenarios.
Is this example the news request is simple. However what if the request was more complex, what if we needed to add headers etc to that request? All that code would be in this class.
What if we wanted to change the request used to fetch the news? We would have to make a code change here. Or what if we could fetch news from more than one API? How would we do that in the current structure?
Once the request has been made we are using a JSONDecoder to decode the response. What if the response comes back in a different format? What if we wanted to use a different decodable for the response?
What if the news request can be used in multiple places?
As we can see from the above list, there are several scenarios that would require a code change of this class. If we recall what the single responsibility stands for:
A class should have only one reason to change
There is also a side effect of this which isn’t as obvious, that is testability. Let’s look at some examples:
How would we test changes the URLRequest? If did indeed change the URLRequest or it was being generated differently, how would we test that?
How do we test how our class handles responses from the server? What happens if we get an error for example?
How do we test our decoding code? How can we be sure that it is going to handle incorrect data correctly? How does our news datasource handle decoding errors?
If we look at the code in the example we can see that in would be impossible to write unit tests covering any of the above scenarios. Let’s have a look at how we can break this class down into single components, allowing us to make changes only in one place and at the same time improving testability.
Breaking it down
URL Builder
Let’s start by breaking this class down into separate classes, each with one responsibility. First of all let’s take out the building of the URLRequest and put it in another class.
Great, now we have a class that’s only responsibility is to build and return a URLRequest. In a more complex system this class might need ids, user tokens etc in order to configure the request. In the scenario where we need to change how news is retrieved we only need to modify this one class in order to make that change. We can also change the hostname based on the environment such as dev, test and prod.
The other benefit of doing this is we can now write unit tests to make sure that the URLRequest is being built correctly. Let’s do a small example now:
classURLBuilderTests:XCTestCase{functestURLBuilder()throws{letbuilder=NewsURLBuilder(hostName:"http://mytest.com/")letrequest=builder.getNews()XCTAssertEqual(request.url?.absoluteString,"http://mytest.com/SomeNews/URL","Request URL string is incorrect")}}
Our URL builder isn’t particularly complex so doesn’t need many tests. But at least here with it being in a separate component we can test the construction and make sure it’s being created correctly. We could expand this test to test other elements of the request if needed, or if we needed different parameters to build the request.
Parser
Next lets take the parser and put that into it’s own class.
Here we can see we have taken our decoding code and put it into a separate class. This class has one reason to change, it only needs to be changed if the parsing needs to be changed! Also like our URL builder class we can now test the decoding to make sure we get the results we are expecting:
Now if we look here we can see that we have swapped out the code for building the request and parsing the data for our separate classes. Now our example is following the single responsibility principle. We have 3 components now:
A component to build our request
A component to execute the request
A component to parse the data we get back from the request
So what have we gained:
We now have test coverage of our components (we could update the NewsDatasource to have tests too but that is a bit more advanced and out of scope of this article)
We have the ability to re-use these components in other parts of the app or in other apps if we need to
If we need to make changes, each component is only responsibility for one thing, so we can update and test each change in turn. Rather than making multiple changes in one place and not be able to test them!
As I’m sure any iOS developer now knows, the future of iOS app development is SwiftUI. Apple’s new UI development language is now on its 2nd major release. While my own personal feeling is that the framework is not quite ready for prime time (much like when Swift first arrived. It’s missing some fairly key features) and we are perhaps a version or 2 away from it realistically being an option for being used to build a complete app. There is no denying that it is the future and when it works well, it makes building UI a lot easier.
If you are using an advanced iOS architecture VIPER, MVVM etc it is probably the case that you have abstracted your routing or creating of your view controllers away into a separate part of your architecture. When you need to navigate or deal with a deeplink for example, you will hopefully have something like a dependency injection framework or factory to create your view controller. This is than pushed onto the navigation stack or presented by the view controller depending on what you are trying to do.
This is something that is fairly straight forward in UIKit and makes a lot of sense. In this article we are going to discuss the workings of SwiftUI and how that approach is no longer possible.
This is a fairly simple SwiftUI example but let’s talk through it.
First we have a ContentView, this contains a NavigationView which is kind of similar to a UINavigationController in UIKit, however it is a lot more limited. We have a navigation link that allows the user to tap the text view and will ‘push’ the detail view on to the stack.
Second we have our detail view that simply displays some text.
If we run the code and tap on the links we should get something like this:
Seems to work as intended right? What problems are there with this approach?
There is a tight coupling between the ContentView and the DetailView, what if we want to change the navigation destination?
What if we want to use the ContentView in a different app that doesn’t contain a DetailView but something else?
What if the DetailView has dependencies we need to inject when it’s created? How does the ContentView know what to do in order to create the DetailView?
What if we wish to perform an event such as fire an analytics event before moving to the next view?
What if we wish to present the view in a modal rather than pushing it to the navigation stack?
Many of the more advanced architectures and frameworks have already solved these problems using a router / co-ordinator pattern. These are responsible for handling any navigation logic and often talk to a dependency injection module in order to create the destination view and pushing it onto the navigation stack or presenting it.
Decoupling the Views
The first thing we can try to do is abstract away the creation of the detail view. This will at least give us the opportunity to change the destination without the knowledge of the ContentView.
First of all we have tried to separate out the creation of the destination view into another object. Ideally we could put this into a protocol but for the purpose of simplicity we have just used an object.
We are injecting the presenter into the ContentView now, you will also notice in the NavigationLink we are now calling a method on the presenter to get the destination.
What does this give us that the previous example doesn’t?
There is no longer tight coupling between the ContentView and the DetailView. The destination is no longer hardcoded. If we make the presenter using a protocol for example. We can inject different presenters and have different destinations.
If the detailview has its own dependencies that need injecting then the presenter can take care of that as well without having to code them in here.
However it’s not all tea and biscuits! There are still a number of issues highlighted earlier that this solution doesn’t solve:
We are still not able to trigger any analytics events or any other app behaviours off the back of the navigation trigger. Block the user from navigating until they have logged in for example.
We can’t change or configure how the navigation happens, for example presenting a login vs actually performing navigation.
We are also exposing navigation to the view, a presenter typically would not need to expose navigation functionality to the view. It would handle a tap event and then hand off that navigation logic to the router. Here we have to expose that functionality to the view itself.
Keep with UIKit for navigation, for now
My personal feeling is that navigation in SwiftUI could do with some more work. Views themselves should not know or care about where they are navigating to and how. They should be a visual representation of state. Of course, the navigation could be a presentation of state too, however a quick peak at the NavigationView docs shows no access to any form of state at all. The navigation view polices its own state, nothing outside of the object has a way to modify that state.
Further to that, many of the methods we have come to expect from UINavigationController are simply not available here. Whether it’s lack of maturity or a slightly confused approach I don’t know. My recommendation for now would be to make use of UINavigationControllers and the UIHostingController to perform navigation for the time being, at least until a better way to manage and manipulate the navigation state is added to SwiftUI.
Let’s have a quick look at how that changes things. First we need to create a hosting controller and inject our SwiftUI view:
So here we are creating our presenter and our view as before but adding them into a UIHostingViewController and a navigation controller. The UIHostingViewController allows us to put SwiftUI views into what is essentially a UIViewController and use it within a UIKit app.
We have also passed a reference to the navigation controller to the presenter. Let’s have a look at our updated SwiftUI code now that we have refactored it into a UIHostingController.
// 1finalclassContentPresenter:ObservableObject{weakvarnavigationController:UINavigationController?funcbuttonTapped(){// Do whatever we like// ...// Navigateletvc=UIHostingController(rootView:DetailView())navigationController?.pushViewController(vc,animated:true)}}// 2structContentView:View{@ObservedObjectprivatevarpresenter:ContentPresenterinit(presenter:ContentPresenter){self.presenter=presenter}varbody:someView{Button(action:{presenter.buttonTapped()}){Text("Navigate")}}}
What’s changed here:
First of all our presenter has replaced our getDetailView with a more generic button tapped function. This function can do any number of things we need it to do, before finally navigating. Here you can see we are using our reference to the navigation controller to push the new view controller.
In our SwiftUI view you will see we no longer have a NavigationView or a NavigationLink. Our view has become far more generic and doesn’t contain and navigation specific logic. You will also see that we have a button which has a tap action assigned by the presenter. This allows us to make the button do anything, not just trigger navigation.
Hopefully you found this helpful when exploring navigation options in SwiftUI. You can find the SwiftUI Sample and the UIHostingController samples on github.
As I’m sure any iOS developer now knows, the future of iOS app development is SwiftUI. Apple’s new UI development language is now on it’s 2nd major release. While my own personal feeling is that the framework is not quite ready for prime time (much like when Swift first arrived. It’s missing some fairly key features) and we are perhaps a version or 2 away from it realistically being an option for being used to build a complete app. There is no denying that it is the future and when it works well, it makes building UI a lot easier.
As SwiftUI is the future, I’ve been investigating how teams might migrate their existing architectures across to the new technology. There a number of challenges presented by migrating to SwiftUI we will discuss below. As the title suggests we will be exploring how to use a presenter to control a SwiftUI view. It doesn’t matter which architecture you are using as such, whether it’s VIPER, MVVVM, VIP, MVP etc. As long as the logic and state of the view has been abstracted away from the view itself so it can be properly unit tested.
Example
List Item View
Let’s start by creating an example in SwiftUI. We will create a simple list view to display some news for example. Let’s create a simple list view first of all:
This is quite a straight forward view, but let’s step through it.
First of all we define our model for the view. We have an id so that we can conform to Identifiable. This allows SwiftUI to uniquely identify each model in the view and helps with performing things like animation and reordering. We also have a title, optional subTitle and an image string. Hopefully nothing here is too scary.
Now we define the view inself. Views in SwiftUI are simple structs that conform to the View protocol, rather than subclasses of UIView like they used to be in UIKit. Its a simple Hstack with an image view then 2 labels stacked on top of each other. See the screen grab below.
Finally we have the preview code to inject an example model to use in the preview.
List View
Now that we have the items in our list, lets create a simple list view that displays those items.
A simple ContentView who has an array of list item view models and a body. The body lists out the content of our list using the ListItemView we created earlier. Simple
Here we have some test data to show that our list is working. If we preview this view we will see something like this:
That’s wonderful, however it is not particularly dynamic. This is where a presenter or view model would come in. If we look at the description of MVP or MVVM we will see they have a similar role:
The presenter acts upon the model and the view. It retrieves data from repositories (the model), and formats it for display in the view.
There are further abstractions layers (such as interactors and use cases). However we are less concerned with them in this discussion and more on the relationship between the view and the presenter who is holding the state and logic of the view.
Abstracting the state
So at the moment we have a pretty stateless SwiftUI view that simply displays a hardcoded list. Now let’s attempt to abstract the list items away into another object that is injected into the view. This object would be responsible for fetching our items and loading them for the view.
When your type conforms to ObservableObject, you are creating a new source of truth and teaching SwiftUI how to react to changes. In other words, you are defining the data that a view needs to render its UI and perform its logic. SwiftUI uses this dependency to automatically keep your view consistent and show the correct representation of your data. We like to think of ObservableObject as your data dependency surface. This is the part of your model that exposes data to your view, but it’s not necessarily the full model.
So lets update our example to move our list of items into a separate class that conforms to this protocol.
The @Published property wrapper here works with the ObservableObject to notify SwiftUI whenever the value changes in order to trigger an update.
We also see the @ObservedObject property wrapper. This causes this view to subscribe to the ObservableObject that’s assigned to it, and invalidates the view any time the object updates.
This is great and allows us to inject an object from outside of the view who can manage the fetching and supplying to the view, whenever the listItems var updates, the view will automatically update!
De-coupling dependencies
Now there is one problem here, can you see it? This view has a dependency between the view itself and the presenter class. Now if you are following the SOLID principles for example, and like to separate dependencies between your classes and layers we will need to remove the dependency between the view and presenter.
To do this lets change the ListPresenter class to be a protocol instead:
This looks like it should be a straight forward change… Wrong! You will now start seeing errors galore. The primary cause coming from the decleration of our new protocol:
Property ‘listItems’ declared inside a protocol cannot have a wrapper
The problem here being exactly as the error states. We cannot use property wrappers inside protocols! That is going to cause a bit of a problem as we now can’t make use of the nice integration with SwiftUI via @Published properties, or so it seems…
Let’s take a step back for a moment, what exactly does the @Published property wrapper actually do? The @Published property wrapper essentially provides a publisher that the SwiftUI system can subscribe to in order to listen to updates to the value. This is in fact an implementation detail. One of the key points of protocol oriented programming is to abstract the implementation of functions are variables away from the dependency so that it is unaware of the inner workings. By trying to apply a property wrapper to the protocol we are trying to enforce how that variable should implemented under the hood. When infact should the implementing class of our protocol wish to, they could create their own custom implementation of the wrapper.
Fixing the errors
Ok so let’s start by removing the @Published property wrapper from our protocol:
Great! However there are now a bunch of different errors occuring… The key one that we need to pay attention to is this one:
Protocol ‘ListPresenter’ can only be used as a generic constraint because it has Self or associated type requirements
Right, so we have solved the riddle of @Published but this has now surfaced another problem. In order for our ListPresenter protocol to be compatible with the ObervedObject property wrapper in the view, it must extend ObservableObject. Now the problem here is that the ObservableObject uses an associatedtype. Which means if we wish to use it or hold a reference to it we must do type erasure (for more info read my previous post on type erasure) or use a generic constraint.
The simplest solution is for us to use a generic constraint on the view. View the code below:
So what has changed here. You will now notice that we have added a generic type T to our view. We have also added a generic constraint when implementing the View protocol which signals that the init and body implementations here are only when type T is a ListPresenter. Now in this instance that works fine as we only intend to use this view with our ListPresenter class. This removes the errors and the code now compiles. Let’s update the code and run a little test to make sure we are still getting all the reactive goodness of SwiftUI.
We have updated our list presenter implementation class to update our list items after 5 seconds. Nice and easy. If we initialise our view with a presenter with 5 items as below, then after 5 seconds our list should reduce to the 2 items as set in the timer.
Now let’s run this as part of an app and see what happens:
So as you can see, after 5 seconds the list of items is reduced after 5 seconds to 2 items, proving that our implementation works and we are still able to hook into the nice secret sauce that combine and swiftUI expose to us to allow us to update our views. I’ve seen some rather crazy implementations and workarounds on Stack Overflow. Hopefully this implementation is a little nicer!
Download the sample project to run it for yourself (Xcode 12.4)
All apps developed require data of some description. This data is stored somewhere, could be on the device itself, in a remote database/service or a combination. Let’s take a look at the most common sources of data:
Each of these methods saves data in a different format. Now I’m sure you will have used at least one of these methods in your apps at some point to retrieve / save data.
When not using the repository pattern it is quite common to access and use these elements directly, either in your ViewController or in some other part of your app depending how it is structured.
The problem
What’s the problem with this approach? Your app becomes difficult to maintain. Now if you only have a small app with a few screens then this isn’t much of a problem as there are only a few elements to change.
However, what if you are working on a large app with several developers and lots of code? You could have NSManagedObjects or Codable objects littered throughout the codebase for example. What happens if you wish to remove Core Data? Perhaps move to realm? You would need to modify all of the classes in your codebase where you had used your Core Data objects.
Similarly, if you are using Codable objects directly from your JSON response. What happens when your backend team changes the API or you switch to a different API provider? The structure of the data may change which means your Codable objects might change. Again you will need to modify a large number of classes if you are working on a large app.
We can also apply this to the other options such as accessing data from 3rd party frameworks. If we use the objects returned from the framework directly, they will all need changing if we change provider or the SDK changes.
There is also the question of query language. Web services use headers and URLQueryItem, Core Data uses Predicates and so on. Every entry point to query the data must know and understand the underlying query language in order to get the information it once. Again, if this changes we need change every query point to the new format.
Let’s have a look at the diagram below:
Here we have an app structure that is making use of Core Data. There is an object that is being used to access the stack that returns some data. Let’s say for this example that it is news articles. These new articles must inherit from NSManagedObject to be used in Core Data. Now if our data layer is returning NSManagedObjects to the rest of our app structure we now have a dependency between Core Data and the rest of the files in our app. If we wish to move to Realm for example, or switch to using some other form of data store we would need to modify all the of files in the app. The app in this example is only small, imagine having to do that for a much bigger app!
Domain Objects and the Repository
This is where Domain Objects come in. Domain Objects are value objects that are defined by your application. Rather than using objects and structures defined outside of the app, we define what we want the objects to look like. It’s then up to the repository to map between the data storage object / structure to these value objects.
When we do this, it means any changes to the data access layer, as we discussed earlier such as data structure changes or changes in provider don’t impact the rest of the app. The only part of the app that needs to be updated is the repository and it’s mapping to the domain objects.
The below quote summarises the idea of the pattern:
Repositories are classes or components that encapsulate the logic required to access data sources. They centralize common data access functionality, providing better maintainability and decoupling the infrastructure or technology used to access databases from the domain model layer.
Let’s have a look at our previous example but modified to use the a repository and domain objects:
So what is the difference here? As you can see the Core Data stack is still returning NSManagedObjects, however the repository is converting that to a domain object. This object doesn’t inherit from NSManagedObject, also it’s structure and attributes are defined by the app rather than what is in the data store.
Now if we wanted to move away from Core Data to something else the only classes that need to be changed are the Core Data stack and the repository. The rest of the app does not need to be changed as we can map the new data stores type to our domain objects using the repository.
Example
To show a small working example we are going to use a couple of Free Public APIs (highly recommend this resource if you are looking to build a demo app or experiment). We will use 2 APIs that returns users. However they return them in a different format.
As we have done in previous blog posts we are going to use QuickType to generate our Codable objects from our JSON response. We will start with our first request.
This structure will allow us to decode the response from the first request. Let’s make a simple example that takes the response and outputs some data. We will be using code from our Simple JSON Decoder to process the output so feel free to read up if the code you see doesn’t make sense.
First of all we are making the request using our Simple JSON Decoder to return our new User type.
Output any errors
So here we are outputting the name, address and location of the user we get back. Super simple right now.
Managing change
Now let’s say we change provider. Maybe our backend team changes the API, or we switch data provider or from 2 different data provider SDKs. In our example we will switch from the first url (https://jsonplaceholder.typicode.com/users/1) to the second (https://randomuser.me/api/).
The first thing we will need to do is change all of our codable objects as the structure of the response is different. Let’s use QuickType again to give us the new structure:
Now this is more complicated that it needs to be for our example but I’m leaving it here as an extreme example of how different things can be. As you can probably tell the structure and types have change dramatically from our first example. So let’s try and output the same data from this example in our previous example. We can ignore the request part and just focus on the data output so we can see the differences:
As you can see from even this simple example. We would have to change 7 lines of code, just to produce the same output. Now imagine this change happening on a much bigger project! There could possibly be 100s of lines of code that would need updating, all because the API response has changed.
Repository Pattern
Here is where the repository pattern comes in. We can create a user repository that fetches the user and converts it to our domain object. That way we don’t need to update the output.
First thing to do is design our domain object that will represent a User in our system. Now all we are doing in this simple example is outputting a few attributes so let’s design our object with just those attributes as we don’t need the rest.
Here we have a nice simple representation of our User object. There is no need to consider any of the other possible attributes returned from the API. We aren’t using them in our application and they will just sit around taking up valuable memory. You will also notice that this object doesn’t conform to Codable or subclass NSManagedObject. This is because DomainObject should not contain any knowledge about how they are stored. That is the responsibility of the repository.
To design our repository we can make use of Generics and Protocols to design a repository we can use for anything, not just our DomainUser. Let take a look:
Here we have different functions for all of the operations we can do. What you will notice is that none of these functions specify where or how the data is stored. Remember when we talked about different storage options at the beginning? We could implement a repo that talks to an API (like in our example), one that stores things in Core Data or one that writes to UserDefaults. It’s up to the repository that implements the protocol to decide these details, all we care about is that we can load and save the data from somewhere.
See it action
Now we have defined what the repository pattern is, let’s create 2 implementations. One for our first request and one for the second. Both should return domain objects, rather than the type returned from the request.
Here is our example from earlier in the article but updated to use our new repositories. Here we go and fetch the user and print their details, the same as before. Now below we can switch to our second request and see how that will work.
Now notice how the only part we changed was the implementation class? The rest of the code remained the same even though where the data was coming from has changed and is coming back in a completely different structure. Now imagine we are using this repo in many places to fetch user details. We can quickly switch between different data sources without changing the code that uses it. The only changes we have to make are to the repo and to the data mapping code. So only one change rather than a change in every single class that uses these objects.
Conclusion
So let’s recap what we have discussed here:
First of all we discussed the problem of using data storage classes throughout your codebase. Especially on large projects if you need to switch data source / structure.
We then discussed how using the repository pattern and mapping to domain objects rather than using data storage classes can make your code easier to change in the future.
We worked through some examples of how changing API structures can impact your code.
We then implemented a basic repository pattern with mapping to domain objects to show how doing this can make updating your project easier.
Finally let’s discuss the pros and cons of the approach:
Advantages
Code is easier to change if you need to switch data source or structure
Separates concerns of where / how data is stored away from the rest of your app
Disadvantages
Adds more code and complexity
Need to write mappers for each object to domain objects
Following on from the previous post where we explored simple JSON decoding. In this post we are going to extend that simple networking example to include some offline caching so that if our network requests fail we can still provide content.
I hear of a lot of people put off by providing a caching layer in their apps for data. Popular solutions include putting an actual relational database inside your app to cache the data such as using Core Data or Realm. Now these solutions are fine if you are intending to levarage the power of a relational database to perform some kind of task. However they add a lot more complexity if you are simply using them as a caching layer. A few draw backs below:
If you are consuming in an house API you may be trying to replicate a back end database structure driven by a much more capable server side DBMS.
Whether you are mapping against a back end database or just the returned JSON. What happens when the structure changes? Do you need to update your parsing and data structure. Do you need to migrate data?
Not only do you need to make sure you can parse and store the data correctly, you then must make sure you query the data in the same way so that you get the same results as returned from the API.
What happens if multiple requests need to update the data? Handling concurrent data updates carries it’s own complexity and head aches.
This is just a sample of the challenges faced when trying to use a relational database as a caching layer for your app. Now you could build some custom caching layer that writes things to disk or a library such as PINCache. However what if I told you there is something simpler that is already built in to iOS as standard?
Caching Headers
To help explain this we need to explore how HTTP caching headers are intended to work. Now, most request to an API will return a bunch of HTTP headers. These provide information and advice to the receiver about the response to the request. We won’t cover them all but the one we are most interested in for this example is the Cache-Control header.
Cache-control is an HTTP header used to specify browser caching policies in both client requests and server responses. Policies include how a resource is cached, where it’s cached and its maximum age before expiring (i.e., time to live)
The part of this statement that talks about maximum age before expiring is what we will explore here. Most APIs will specify something called a max-age in the response headers. This is the length of time in seconds that the receiver should consider this information valid for. After that period of time the response should be considered stale and new data be fetched from the source.
By default URLSession and URLRequest have a cache policy of useProtocolCachePolicy. This means they will honour the HTTP caching headers when making requests. In the case of the above it will cache request responses for the time specified in the header. It is possible to override this behaviour if you wish using one of the other options.
Postman
To demonstrate this behaviour in action we are going to use a tool called Postman. You may be using this already, it’s a fantastic tool for developing and testing APIs. One of the services that are provided by Postman is something called Postman Echo. This is a service that allows you to send various URL parameters to it and have postman reply those items back to you in a certain response format. To test our example we are going to use the response headers service that is provided, this allows us to specify headers and values in the url query string and have them played back to us in the actual response headers.
If we hit the URL below, you will get a response with the specified headers that you send in the URL query params.
We get back a header in the response of Cache-Control: max-age=30. This means that anyone processing the response should cache the response for 30 seconds at the most before requesting fresh data, as discussed previously.
We can use this to prove how the caching works in URLSession.
Caching in URLSession
Let’s setup an example below on how to test out these cache headers:
Let’s step through this example step by step to demonstrate what is happening:
First of all we setup our postman echo request, we will set this up to return cache headers of 30 seconds
We make a request using the standard dataTask method. When we get the response we cast it to an HTTPURLResponse. An HTTPURLResponse contains a dictionary called allHeaders which is a dictionary containing all of the headers returned in the response. However this is error prone as dictionary keys are case sensitive. To combat this Apple have added a new function called value that takes a key string but does a case-insensitive match against the header so you don’t have to.
With the code in point 2 we are grabbing the date of the response and the cache control header and printing them to the console so we can see what they are.
We sleep for 5 seconds then perform another request.
Here are performing the same request as above and fetching the values of the response headers again. This will help us to see how the caching works.
If we run the above code in our playground we should see the following in the console:
The first 2 lines show that our request executed at a certain date and time, the second line displays the cache header we configured in our postman echo request.
The last 2 lines show the same thing?
This is because we set a cache time of 30 seconds in the request header. As you know from step 4 above, we slept for 5 seconds inbetween each request. The fact the date headers are the same shows that the second request response is in fact the same response as the first request, it has just been fetched from the cache.
To prove this we can modify the request so that we only cache the response for 3 seconds, this way when we sleep for 5 seconds, the response from the first request should be considered stale and the second request should end up with a new response.
Let’s modify the URL in our request to set the cache control to 3:
How is this different from above. The main difference you will notice is that the request times are now different. The second request timestamp is 5 seconds after the first. This is because our cache time is 3 seconds now, so the second request is no longer pulling from the cache and is in fact a new request with a new response.
Offline and the Network Conditioner
Now you are probably asking yourself what this has to do with offline caching? To understand how we can leverage this caching behaviour we need to throttle our requests so that they fail. One of the tools at our disposal is something called the Network Conditioner. This is provided by Apple in the additional tools download.
If you download the tools and install the network conditioner preference pane, you should be able to launch it from your Mac preferences. Once open you should see something like the below:
This tool allows you to create various network conditions on your mac, such as 100% packet loss, 3G, dial up etc. We are going to use this to replicate a connection failure in our example to see how we can begin to use some of URLSession’s properties to access cached request data.
If we add the below into our second request callback so we can see if the request errors:
If we run the sample again, however this time once we receive the first 2 console messages. We activate the network conditioner using 100% loss. This will cause the second request to fail (it may take a few seconds for the request to timeout).
If done correctly we should see something like below in the console:
Request1 date: Tue, 23 Jun 2020 12:36:09 GMT
Request1 Cache Header: max-age=3
The request timed out.
Now we aren’t getting a response from the second request. Instead we are receiving an error. This is expected behaviour as the second request is indeed failing. What we can do in this scenario is grab the response from the cache if we so wish. To do so add the code below to the second completion handler:
Part of the URLSession is the URLSessionConfiguration. This is an object which provides all of the configuration for the URLSession. One of the attributes here is the URLCache. This is where the magic happens. It is an in-memory and on-disk cache of the responses from URLRequests. It is responsible for storing and deleting the data, in our example that is controlled via the response headers.
One of the methods on the URLCache is cachedResponse. This will return the cached response for any URL request still in the cache.
In the example above we are pulling the cached response and outputting the header of the attached HTTPURLResponse. If we take another look at our example with the additional 2 snippets above we should have something like below:
We have created a method the same as the standard dataTask method on URLSession. However we have added a bool to control whether we would like to return the cached response on receiving a network error.
Here we take the example we used earlier in our example and applied it to this method. First we check whether we should return the cached response based on our cachedResponseOnError parameter. Then check to see if we do have an error, if we do then attempt to grab the cached response from the URLCache and return it’s data and response objects along with the error.
In the case where any of the above fails we simply return everything exactly as it was in returned by the normal dataTask method.
As the completion handler returns Data, URLResponse and Error we are able to return the data and response even if there is an error. That is a bit of a disadvantage in this case as the function caller needs to be aware that they may receive an error but also the cached response as well so need to cater for those scenario themselves.
Combine
Hopefully you have at least heard of Combine even if you haven’t had chance to use it yet in a production app. It is Apple’s own version of a reactive framework. Those of you who have already been using RxSwift will be right at home. We aren’t going to go into too much detail about what Combine is but here is a definition of what reactive programming is:
In computing, reactive programming is a declarative programming paradigm concerned with data streams and the propagation of change
In more simplistic terms, reactive programming uses a observer pattern to allow classes to monitor different streams of data or state. When this state changes it emits an event with the new value which can trigger other streams to perform work or update such as UI code. If you are familier with KVO you will understand the basic concept. However reactive programming is far less painful and a lot more powerful than KVO.
Now the approach described in the previous section works fine in the case of the completionHandler as it allows us to return all 3 items to the caller regardless of what happens. However in combine streams are designed to either return a value OR an error.
First I am creating a typealias of the dataTaskPublisher.Output. This is mostly for code formatting reasons as the string is very long. This is simply a Data object and a URLResponse object in a tuple.
Here we have setup our function with the cachedResponseOnError flag the same as before. We are returning a publisher with our type aliased output.
First we call the standard dataTaskPublisher method to setup our request. We immediately chain that publisher using the new tryCatch. So what does this do? “Handles errors from an upstream publisher by either replacing it with another publisher or throwing a new error.” So here we catch any connection errors from the dataTaskPublisher and we can either throw another error or send another publisher down the chain.
So the same as our pure Swift example we attempt to fetch our response from the cache, except here if we fail to find anything in the cache we just rethrow the same error we received in the try catch block so it can be handled further down the stream.
If we were able to find a response in the cache then we use the Just publisher to send the new value down the stream wrapping our cached response.
We erase the types to AnyPublisher using type erasure so items further down the stream don’t need to know the types. For more info on type erasure see my previous post.
Now let’s take our previous test and adapt it so we can see this in action:
As we have done with our other examples let’s step through and see what happens.
First we create our request as we did in the pure Swift example, and then create our new publisher using our newly created function.
Now in Combine, publisher’s only execute once there is an unsatisfied subscription. This happens whenever the sink function is called. The first closure is called when either the stream completes or an error throws (which also terminates the stream). The second closure is called whenever a new value is published from the stream. In this first case we can a tuple containing Data and a URLResponse. As before we inspect the date header of the request.
As before we sleep for 5 seconds (we have set a timeout of 3 on using the cache control headers)
This matches the same as step 2 however we changed the output.
If we follow the same steps as before and turn on packet loss using the network conditioner when we hit step 3 we should see a console log like below:
Now, this proves that we are returning our cached response in the second request as we have no connection and we are still receiving a response. However what is the problem here?
There is no error
The above implementation works fine if we just want to display cached data. However you may wish to inform the user that there was a connection failure and they are viewing old / stale information. So how can we get around this? In order to send the cached value, then the error we would need to create a custom Combine publisher. We won’t cover that here as that is a post in itself.
Conclusion
We have shown how we can make use of built in functionality of URLSession and URLCache along with the HTTP standard headers to provide simple and basic offline caching.
Advantages
Simple implementation, doesn’t require 3rd party frameworks or complex relational databases
Makes use of 1st party frameworks and relies on currently available standards (HTTP)
Concurrency handled automatically
Disadvantages
Relies on cache-control headers being correctly used in the API being consumed
URLSession cache policies need to be configured correctly
Doesn’t support more complex caching requirements / rules if needed
URLCache will delete the cache if device memory becomes full, something to bear in mind using this approach
In summary, this approach is simple and provides basic offline functionality for your app. If you have more complex needs / requirements for caching your data then other approaches may be more suitable.