The announcement for the Plex UNO release came yesterday, but for several people across different platforms it had been there in one fashion or another for a while.
If you happen to have an Android TV, Roku, XBOX, etc, you might have noticed that Plex operated differently, had different features, and otherwise was not the same from device to device.
UNO attempts to reverse this wild trend … ok, I’m going to skip all the Uno draw card references here… anyway, UNO attempts to bring all major supported devices into the same layout more or less.
One major difference from some older versions is the Hubs setup… if you only get your Plex content from one server this probably isn’t much of a thing for you, but if you happen to have multiple servers and share content with a lot of people like I do, chances are it’ll be a really really useful feature.
However, as we know, all change is bad and should be feared… rrrr!
It’s actually where I want it more or less. Now if they can just get it so that you can Chromecast live TV from your Plex app and add multiple types of tuners to one Plex server (iptv + OTA but using different channel guides,) I’ll have nothing to gripe about.[Draw 4]
Powered by WPeMatico
- Facebook Patents Shadow Banning — which has a long history elsewhere.
- Living Off The Land in Linux — legitimate functions of Unix binaries that can be abused to break out restricted shells, escalate or maintain elevated privileges, transfer files, spawn bind and reverse shells, and facilitate the other post-exploitation tasks. Interesting to see the surprising functionality built into some utilities.
- Neural Blind Deconvolution Using Deep Priors — deblurring photos with neural nets. Very cool, and they’ve posted code. (via @roadrunning01)
- Warshipping (TechCrunch) — I mail you a package that contains a Wi-Fi sniffer with cellular connection back to me. It ships me your Wi-Fi handshake, I crack it, ship it back, now it joins your network and the game is afoot. (via BoingBoing)
Powered by WPeMatico
As companies begin to explore AI technologies, three areas in particular are garnering a lot of attention: computer vision, natural language applications, and speech technologies. A recent report from the World Intellectual Patent Office (WIPO) found that together these three areas accounted for a majority of patents related to AI: computer vision (49% of all patents), natural language processing (NLP) (14%), and speech (13%).
Companies are awash with unstructured and semi-structured text, and many organizations already have some experience with NLP and text analytics. While fewer companies have infrastructure for collecting and storing images or video, computer vision is an area that many companies are beginning to explore. The rise of deep learning and other techniques have led to startups commercializing computer vision applications in security and compliance, media and advertising, and content creation.
Companies are also exploring speech and voice applications. Recent progress in natural language and speech models have increased accuracy and opened up new applications. Contact centers, sales and customer support, and personal assistants lead the way as far as enterprise speech applications. Voice search, smart speakers, and digital assistants are increasingly prevalent on the consumer side. While far from perfect, the current generation of speech and voice applications work well enough to drive an explosion in voice applications. An early sign of the potential of speech technologies is the growth of voice-driven searches: Comscore estimates that by 2020 about half of all online searches will use voice; Gartner recommends that companies redesign their websites to support both visual and voice search. Additionally, smart speakers are projected to grow by more than 82% from 2018 to 2019, and by the end of the year, the installed base for such devices will exceed 200 million.
Audio content is also exploding, and this new content will need to be searched, mined, and unlocked using speech technologies. For example, according to a recent New York Times article, in the US, “nearly one out of three people listen to at least one podcast every month.” The growth in podcasts isn’t limited to the US: podcasts are growing in other parts of the world, including China.
Voice and conversational applications can be challenging
Unlike text and NLP, or computer vision, where one can pull together simple applications, voice applications—that venture beyond simple voice commands—remain challenging for many organizations. Spoken language tends to be “noisier” than written text. For example, having read many podcast transcripts, we can attest that transcripts from spoken conversations still require a lot of editing. Even if you have access to the best transcription (speech-to-text) technology available, you often end up with a document with sentences that contain pauses, fillers, restarts, interjections (in the case of conversations), and ungrammatical constructs. The transcript may also contain passages that need to be refined due to the possibility that someone is “thinking out loud” or had trouble articulating or formulating specific points. Also, the resulting transcript may not be properly punctuated or capitalized in the right places. Thus, in many applications, post-processing of transcripts will require human editors.
In computer vision (and now in NLP), we are at a stage where data has become at least as important as algorithms. Specifically, pre-trained models have achieved the state-of-the art in several tasks in computer vision and NLP. What about for speech? There are a few reasons why a “one size fits all” speech model hasn’t materialized:
- There are a variety of acoustic environments and background noises: indoor or outdoor, in a car, in a warehouse, or in a home, etc.
- Multiple languages (English, Spanish, Mandarin, etc.) may need to be supported, particularly in situations where speakers use (or mix and match) several languages in the course of conversations.
- The type of application (search, personal assistant, etc.) impacts dialog flow and vocabulary.
- Depending on the level of sophistication of an application, language models and vocabulary will need to be tuned for specific domains and topics. This is also true for text and natural language applications.
Building voice applications
Challenges notwithstanding, as we noted: there is already considerable excitement surrounding speech technologies and voice applications. We haven’t reached the stage where a general-purpose solution can be used to power a wide variety of voice applications, nor do we have voice-enabled intelligent assistants that can handle multiple domains.
There are, however, good building blocks from which one can assemble interesting voice applications. To assist companies that are exploring speech technologies, we assembled the following guidelines:
- Narrow your focus. As we noted, “one size fits all” is not possible with the current generation of speech technologies, so it is best to focus on specific tasks, languages, and domains.
- Understand the goal of the application, then backtrack to the types of techniques that will be needed. If you know the KPIs for your application, this will let you target the language models needed to achieve those metrics for the specific application domain.
- Experiment with “real data and real scenarios.” If you plan to get started by using off-the-shelf models and services, note that it is important to experiment with “real data and real scenarios.” In many cases, your initial test data will not be representative of how users will interact with the system you hope to deploy.
- Acquire labeled examples for each specific task. For example, recognizing the word “cat” in English and “cat” in Mandarin will require different models and different labeled data.
- Develop a data-acquisition strategy to gather appropriate data. Make sure you build a system that can learn as it gathers more data, and an iterative process that fosters ongoing improvement.
- Users of speech applications are concerned about outcomes. Speech models are only as interesting as the insights that can be derived and the actions that are taken using those insights. For example, if a user asks a smart speaker to play a specific song, the only thing that matters to this user is that it plays that exact song.
- Automate workflows. Ideally, the needed lexicon and speech models can be updated without much intervention (from machine learning or speech technology experts).
- Voice applications are complex end-to-end systems, so optimize when possible. Speech recognition systems alone are comprised of several building blocks which we described in a previous post. Training and retraining models can be expensive. Depending on the application and setup, latency and continuous connectivity can be important considerations.
From NLU to SLU
We are still in the early stages for voice applications in the enterprise. The past 12 months have seen rapid progress in pre-trained natural language models that set records across multiple NLP benchmarks. Developers are beginning to take these language models and tune them for specific domains and applications.
Speech adds another level of complexity—beyond natural language understanding (NLU)—to AI applications. Spoken language understanding (SLU) requires the ability to extract meaning from speech utterances. While SLU is not yet on hand for voice or speech applications, the good news is that one can already build simple, narrowly focused voice applications using existing models. To find the right use cases, companies will need to understand the limitations of current technologies and algorithms.
In the meantime, we’ll proceed in stages. As Alan Nichol noted in a post focused on text-based applications, “Chatbots are just the first step in the journey to achieve true AI assistants and autonomous organizations.” In the same way, today’s voice applications provide a very early glimpse of what is to come.
Powered by WPeMatico
- From The Depths Of Counterfeit Smartphones — security look at the counterfeit phones. Spoiler: they’re nasty, stay away. Both the Galaxy S10 and iPhone 6 counterfeits we assessed contained malware and rootkits. And that’s the most straightforward nastiness: even if you removed the rootkit they’d still be shocking. In the case of the “iPhone,” further digging revealed that it runs a far older version of Android: Kitkat 4.4.0. Kitkat’s last update came in 2014.
- Linking Art through Human Poses — arXiv paper that finds artwork with matching poses using OpenPose. (via MIT TR)
- A Framework for Content Moderation (Ben Thompson) — pretty good post, tackling why and where the different levels of moderation make sense.
- Fully Remote Attack Surface of the iPhone (Google Project Zero) — very interesting read, showing the detail and dead ends of a security tester. The method […] processes incoming MIME messages, and sends them to specific decoders based on the MIME type. Unfortunately, the implementation did this by appending the MIME type string from an incoming message to the string “decode” and calling the resulting method. This meant that an unintended selector could be called, leading to memory corruption.
Powered by WPeMatico
La Guerra Asimétrica del Dictador Hugo Chávez: Contra Comunicadores Sociales y Medios en el 2007
Eladio Rodulfo Gonzalez, Periodista Venezolano, publica una Investigación contentiva de varios tomos, donde recopila las Agresiones, Violaciones de Derechos Humanos y Asesinato de Comunicadores Sociales en el Ejercicio de su Profesión, por parte de los Gobiernos Dictatoriales de Hugo Chávez y Nicolás Maduro, aqui comparto el Cuarto Tomo: Año 2007
El año 2007 fue fatal para la libertad para la libertad de expresión pues el 27 de mayo, en contra de la opinión nacional e internacional cerró definitivamente a la estación televisiva más antigua del país, Radio Caracas Televisión, apropiándose de los equipos de la estación. Lo propio haría después con Radio Caracas Televisión Internacional por negarse a transmitir las kilométricas y arbitrarias cadenas del dictador.Las amenazas del Tribunal Inquisitorial del régimen, la Comisión Nacional de Telecomunicaciones, dieron luz a la autocensura en los medios audiovisuales para evitar sanciones.El entonces director de Telesur, Andrés Izarra, sugirió el establecimiento de la hegemonía comunicacional.Continuó el impedimento a los medios privados de hacerle cobertura a los eventos de la dictadura.Venezuela se situó entre los tres países latinoamericanos con mayores dificultades para el ejercicio del periodismo.Continuaron las agresiones verbales y físicas contra comunicadores y medios por parte de las fuerzas de seguridad, especialmente la Guardia Nacional y la Policía Nacional.
Powered by WPeMatico