Installation by teamVOID uses industrial robots to perform life drawings alongside human artists:
‘Way to Artist’ has the purpose of rethinking the process of artistic creation through a comparison of robot and human actions. Drawing is assumed to be a creative activity that only humans are capable of. Nowadays, however, the emergence of artificial intelligence has some believing that artwork could be created by robots. In connection with this, the work involves drawings executed by a robot and a human, each with different drawing skills. In the process, it reconsiders the general meaning of the drawing activity.
Whilst this isn’t the first example of this type of setup, it isn’t clear whether the robots have any visual interpretation model, so this could be a metaphorical rather than technical presentation.
Link
Visual experiment from LuluXXX tranforms a music video using lllyasviel’s code designed to colorize black and white manga:
messing around with style2paint : https://github.com/lllyasviel/style2paints a lot of entropy going on. original video : https://www.youtube.com/watch?v=gGjJFRPtdDI Music performed by Twin Peaks cast member Chrysta Bell and David Lynch, written by David Lynch and Dean Hurley.
An online version of Style2paint can be found here
Another smart AR experiment from Zach Lieberman proving Augmented Reality is an interesting creative platform: this one visualizes audio as it is recording and plays back as you follow the path both forwards and backwards:
A post shared by zach lieberman (@zach.lieberman) on Sep 6, 2017 at 5:55am PDT
Quick test recording audio in space and playing back – (video has audio !) #openframeworks
Link
Fractal Friday ॐ Not my photo ॐ
THOPTER_02
Currently being installed in Times Square, Radius Displays have created a shape changing screen made from an array of motorized panels:
A post shared by Radius (@radiusdisplays) on Jun 16, 2017 at 1:12pm PDT
There is very little information to find about the project, other than it is currently installed and testing and officially launching soon - the best bet is to keep an eye on Instagram here
It will make an official debut on August the 8th [Link]
Research from Carnegie Mellon Textiles Lab have put forward a framework to turn 3D model file into a physical knitted object:
We present the first computational approach that can transform 3D meshes, created by traditional modeling programs, directly into instructions for a computer-controlled knitting machine. Knitting machines are able to robustly and repeatably form knitted 3D surfaces from yarn, but have many constraints on what they can fabricate. Given user-defined starting and ending points on an input mesh, our system incrementally builds a helix-free, quad-dominant mesh with uniform edge lengths, runs a tracing procedure over this mesh to generate a knitting path, and schedules the knitting instructions for this path in a way that is compatible with machine constraints. We demonstrate our approach on a wide range of 3D meshes.
More Here
A teenager managed to develop an app that can help diagnose a diabetes-related condition affecting her grandfather.
Kavya Kopparapu’s grandfather lives in India, where there aren’t enough ophthalmologists to help diagnose all of those who could be affected by diabetic retinopathy.
DR is the world’s leading cause of vision loss in people age 20 to 65, according to the International Agency for the Prevention of Blindness, which estimated that 50% of people with diabetes are undiagnosed.
But in the absence of proper doctors, “computers could be used in their place,” Kopparapu, 16, said in a TEDx talk on artificial intelligence.
Alongside her brother and another classmate, she invented Eyeagnosis, a smartphone app that can photograph patients’ eyes and match them to a database of 34,000 retinal scans collected from the National Institute of Health. Read more (8/8/17)
follow @the-future-now
Researchers at the Massachusetts Institute of Technology’s Computer Science and Artificial Intelligence Laboratory are creating a computer system modeled after the human brain to examine photos of food on social media and break them down into recipes.
“Making an AI machine play and be naturally curious is key”
- Mark Sagar, Academy Award-winning AI engineer on the future of AI in animation.
What’s next for AI? We spoke with 30 visionaries in AI to learn where they think the technology is heading.
See the full interview →
Research from Columbia Computer Graphics Group can create textual encryption by minute altering of font characteristics using neural networks:
We introduce FontCode, an information embedding technique for text documents. Provided a text document with specific fonts, our method embeds user-specified information in the text by perturbing the glyphs of text characters while preserving the text content. We devise an algorithm to choose unobtrusive yet machine-recognizable glyph perturbations, leveraging a recently developed generative model that alters the glyphs of each character continuously on a font manifold. We then introduce an algorithm that embeds a user-provided message in the text document and produces an encoded document whose appearance is minimally perturbed from the original document. We also present a glyph recognition method that recovers the embedded information from an encoded document stored as a vector graphic or pixel image, or even on a printed paper. In addition, we introduce a new error-correction coding scheme that rectifies a certain number of recognition errors. Lastly, we demonstrate that our technique enables a wide array of applications, using it as a text document metadata holder, an unobtrusive optical barcode, a cryptographic message embedding scheme, and a text document signature.
More Here