Appreciate the link, and I do understand that the textures are packed together as tightly as possible to make an efficient atlas. However... in previous versions of xcode even though the atlas didn't show the transparent pixels, it was keeping track of the size properly and would return the images back to you exactly as they were before they were put into the atlas (including transparent portions). As I mentioned above I find it strange that when I ask for the size it says 14x14 (not 12x12) but it's rendering at 12x12. Frustrating. I'm hoping it's just a bug like the many other issues I'm running into (ex: generating a physics body from a texture apparently creates garbage). Appreciate the response(s)!
Topic:
Programming Languages
SubTopic:
Swift
Tags: