A bit of an annoyance when trying to implement geographic region monitoring.
Filed as enhancement request as rdar://16924908 and on Open Radar.
If you like storyboards you might be wanting to use them for larger apps too. But there is source of frustration for us developers. If you have prototype cells that are identical on multiple table view controllers you cannot easily copy them. I filed a feature suggestion for adding global prototype cells.
Filed as rdar://16754279 and on Open Radar.
There’s been a bit of movement on our open source replacement for genstrings. So I figured it would finally be deserving of getting the same treatment as all our other popular repos: automated unit testing, coverage tracking, releases via CocoaPods.
Because of the unwieldy name I refer to it as genstrings2 and it provides the scanning functionally for the Linguan Mac app which was acquired by Peer Assembly.
You know the story: you have an app laid out as storyboard, want to show a modal view controller and when the user is done with it it you want to dismiss it, … or un-present it, or whatever you need to get rid of the modal VC and return. Now what about a view controller that can be both presented modally or via push onto a navigation controller?
Then it starts to get complicated. Unless you know about how to unwind. It’s actually quite awesome, but you have to use it several times because it is a little bit counter-intuitive how to implement unwinding. I’m writing this tutorial so that I myself can remember how to do it in the future.
My publisher Manning is highlighting 3 of their iOS titles on April 11, all three are the “Deal of the Day” meaning that you get 50% off with promo code “dotd041114au” if you purchase them on this day.
Readers of my book get good news today! My publisher Manning released version of the MEAP-version of the book.
Previously available chapters 1 through 3 have been reworked based on the feedback several reviewers gave. There were some tough requests which I just couldn’t fit into the style of the book. But for the most part the book has become much better for it.
If you haven’t pre-ordered the book via Manning’s Early Access Program (MEAP) then please do so today!
Apple’s iPhone 5S features a dedicated motion coprocessor, dubbed the “M7”. Probably because the M stands for Movement/Motion similar to the A in “A7” standing for … um, Apple. This coprocessor collects and analyzes data from multiple iPhone sensors and determines two interesting pieces of information:
Let’s build a functional pedometer app so that we can learn about and explore the latter.
For my book I wanted to see if there is a way to see the performance impact of various settings. So I launched the Energy Diagnostics instrument for the first time. I was not able to conclusively demonstrate the impact of any capture setting, but I did see something odd. Even though my barcode scanner sample app does not have any need for audio data AV Foundation appears to insist on capturing and processing it.
As the energy diagnostics performance instrument indicates, this causes energy usage for “Audio Processing” even though for the purpose of scanning barcodes audio is not needed. In the least developers should be able to indicate to iOS that they have no use for audio processing so that the system can conserve this energy.
This bug report is filed as rdar://16389364 and on Open Radar.