Media Search:



Social media takes on Rush Limbaugh

Social networking sites like Facebook and Twitter lit up this week and conservative commentator Rush Limbaugh is feeling the effects.

Soon after Limbaugh spent hours on his talk-radio show bashing a Georgetown University law student over her support of birth control, people from around the U.S. took to Facebook, Google+ and Twitter to vent their anger at Limbaugh. With no organized leadership, they en masse called for sponsors of Limbaugh's radio show to pull their advertising.

Today, in part because of this social pressure, Limbaugh's show has lost more than 20 advertisers, according to reports, including Allstate Insurance, AOL, Citrix, Quicken Loans and Sears.

Social networking sites are quickly being seen as the new medium for protesters of all types.

"Social media has entered into a new era, which is way past the point when it was used to find the best dry cleaner or see if a certain restaurant had fresh seafood," said Dan Olds, an analyst with The Gabriel Consulting Group. "Now it's become a real tool that is being used to topple governments and even saves lives."

The latest online brouhaha began Feb. 23 when law student Sandra Fluke testified before a congressional committee about the need for insurance companies to cover birth control costs. Limbaugh then called the woman a "slut" and "prostitute" and spent hours discussing her on his radio show.

Word of Limbaugh's actions quickly spread over traditional media and social networking sites, where people were quick to voice their own support or outrage.

For instance, Andy Borowitz tweeted, "Due to remarks of his we consider unacceptable, we have terminated our relationship with Rush Limbaugh. - Satan."

And Stop Rush Limbaugh tweeted, "Please urge Capital One (@AskCapitalOne) to cease advertising on Rush Limbaugh's radio program."

Sensa Products LLC, a weight-loss system company that had been a sponsor of Limbaugh's radio show tweeted Monday: "Rush Limbaugh's comments are not in line with SENSA values so we are pulling our ads indefinitely which shud be down in the next couple days."

Originally posted here:
Social media takes on Rush Limbaugh

Obama press conference: Above all, do no (message) harm

President Obama had a simple goal in his first press conference in more than five months: Dont step on his own message(s).

President Barack Obama gestures during a news conference in the James Brady Press Briefing Room of the White House in Washington, Tuesday, March 6, 2012. (AP Photo/Charles Dharapak)It was clear from the get-go that Obama wanted to accomplish two things with the press conference: First, introduce a series of new housing measures designed to address the foreclosure crisis and second, hammer away at the Republican presidential candidates for their casual (his word) approach to the use of military force.

The challenge for Obama was that he faced a barrage of questions on a series of controversial topics Iran, Israel, Syria, Rush Limbaugh that had the potential to steal headlines from the presidents preferred messaging if he made any sort of news on any of those fronts.

With one possible exception Do you think a president of the United States going into re-election wants gas prices higher? Obama joked at one point in an off-key moment he largely accomplished that goal, virtually ensuring that his tough talk against the GOP candidates on foreign policy matters would be the story of the presser.

Whats said on the campaign trail....those folks dont have a lot of responsibilities, Obama said at one point. They arent commander in chief. At another, he argued: If some of these folks think its time to launch a war, they should say that.

With his rhetoric, Obama was working to make the case that while its easy to call for military action when you are running for president, its much harder to justify it when you are the president and your decision means American troops will be put in harms way.

This was another adult in the room moment for Obama in which he tries to cast himself as someone committed to getting it right while portraying his Republican opponents as interested solely in scoring political points.

(Need more evidence? This is not a game, Obama said at one point regarding foreign policy. And, he repeatedly noted he doesnt undertake the use of military force in a casual manner.)

Even as Obama was trying to make that adult case, he was dodging political pitfalls left and right.

He carefully avoided reiterating his controversial remark that the 1967 boundaries of Israel would be a starting point for negotiations about Middle East peace.

View post:
Obama press conference: Above all, do no (message) harm

Digital Economy Act not in breach of EU laws, Court of Appeal rules

The Court has rejected claims made by BT and TalkTalk that the Digital Economy Act (DEA) violates EU laws. The ruling has been welcomed by the Government and representatives of the creative industries, but may yet be appealed by the ISPs to the UK Supreme Court.

We are pleased the Appeal Court has upheld the original ruling that the Digital Economy Act is a lawful and proportionate response to the threat posed by online piracy," a spokesperson for the Department for Culture, Media and Sport (DCMS) said in a statement.

The DEA is a controversial law that includes provisions aimed at combating online copyright infringement.

In a judicial review ruling in April last year the High Court rejected claims by the ISPs that the DEA violated EU laws.

According to a summary of the Court of Appeal judgment by a coalition of creative industry bodies, the ISPs argued that the DEA breached EU laws on data protection and privacy. BT and TalkTalk also failed to convince the Court that the Act was "incompatible" with provisions set out in the E-Commerce Directive. The Court also rejected claims made by the ISPs that the DEA was unlawful because the Government had failed to give the European Commission enough time to scrutinise parts of the legislation, according to the bodies.

However, the ISPs did successfully argue that they should not be required to pay 25% of the "case fees" that would stem from ISP customers bringing appeals against warning letters they could receive for allegedly infringing copyright under measures allowed for by the DEA. The ISPs will still have to pay 25% of the costs they incur to comply with their obligations under the Act after the Court ruled that it was lawful to impose the charge on them, according to the creative industry bodies' summary.

The measures were always going to have to strike a balance between ISPs and rights holders and the court has decided that the balance struck by the Digital Economy Act is about right, said copyright law expert Iain Connor of Pinsent Masons, the law firm behind Out-Law.com. How this will operate in practice remains to be seen.

"Some may argue that BTs and Talk Talks public stance on the issue left the legislators no room to negotiate and so arguably a more conciliatory approach during the Bills passage through Parliament might have been better or at least so it seems with the benefit of hindsight, Connor said.

Both ISPs said they would consider what action, if any, to take next in light of the ruling.

We have been seeking clarification from the courts that the DEA is consistent with European law, and legally robust in the UK, so that everyone can be confident in how it is implemented," BT said in a statement. Now that the Court has made its decision, we will look at the judgment carefully to understand its implications and consider our next steps."

Continued here:
Digital Economy Act not in breach of EU laws, Court of Appeal rules

Smarter TV: Living Room as Digital Hub From Samsung and Microsoft to Apple and Google

Tim Baxter, President of Samsung Electronics America. Photo by Tim Carmody/Wired.com

NEW YORK Just forget about its giant screen for a moment.

Yes, that new plasma TV is gorgeous, that LED backlight efficient, and that refresh rate ridiculous. But in truth, just like smartphones and tablets, smart TVs are about platforms as much as pictures.

Today in New York, Samsung presented its updated line of smart TVs and related electronics, almost all of them available for sale now. The Korean electronics giant has too many new individual devices from cameras that sync with your TV over Wi-Fi and smartphone speaker docks with honest-to-goodness vacuum tube amplifiers to tricked-out, touchpad-and-microphone-equipped remotes that are 85 to 90 percent of everything you want a smart TV remote to be to give more than passing consideration here to each and every one of them. If you want to get started with that, Ars Technicas Casey Johnston has a great rundown of whats good and bad in all the new interface technologies for Samsungs smart TVs here.

Instead, heres my big takeaway from Samsungs event at least as I see it now, with an eye toward Apples definitely-an-iPad, most-probably-an-Apple-TV event on Wednesday.

In the future, the living room will replace the home office as most households home for the stationary personal computer. Instead of printers and mice and other corded accessories, networked appliances and post-PC machines share data with one another and with the cloud. Play and productivity both become decentered; gaming and entertainment might be on a tablet or a television, with recipes at the refrigerator, a shopping list for the smartphone, and an instructional video on the television set.

All of these experiences will be coherent, continuous and contextual. And like the personal computer at the height of Pax Wintel, the living room will be a platform characterized by triumphant pluralism.

The thing about the living room is that its universal; everyone in the household uses it, Samsung VP Eric Anderson told me at todays event. We know that were not going to capture every single member of the household. In my family, my wife and my daughter are Apple, me and my sons are Android, he noted, pointing out that the majority of devices introduced today can interact with either mobile platform.

The big question for us is what is the core of your household, Anderson added. What is the device of origin? Where do you start, and to where do you return? Thats why we look at the living room, the kitchen along with some mobile devices. Here, no company can be a platform purist: Every consumer electronics company is looking for a differentiator; maybe the differentiator here is the devices ability to work with anything.

Samsungs been manufacturing and selling smart TVs since 2008. Its sets have carried Yahoos widgets, Google TV, and now Samsungs own app-driven Smart Hub software. In those four short years, the technology powering the TV, customers expectations, and entertainment companies willingness to embrace cloud-delivered, app-based over-the-top content have all changed.

Go here to read the rest:
Smarter TV: Living Room as Digital Hub From Samsung and Microsoft to Apple and Google

1953: The Year That Revolutionized Life, Death, and the Digital Bit

Three technological eras began in 1953: thermonuclear weapons, stored-program computers, and modern genetics.

At 10:38 p.m. on March 3, 1953, in a one-story brick building at the end of Olden Lane in Princeton, New Jersey, Italian Norwegian mathematical biologist Nils Aall Barricelli inoculated a 5-kilobyte digital universe with random numbers generated by drawing playing cards from a shuffled deck. "A series of numerical experiments are being made with the aim of verifying the possibility of an evolution similar to that of living organisms taking place in an artificially created universe," he announced.

A digital universe -- whether 5 kilobytes or the entire Internet -- consists of two species of bits: differences in space, and differences in time. Digital computers translate between these two forms of information -- structure and sequence -- according to definite rules. Bits that are embodied as structure (varying in space, invariant across time) we perceive as memory, and bits that are embodied as sequence (varying in time, invariant across space) we perceive as code. Gates are the intersections where bits span both worlds at the moments of transition from one instant to the next.

The term bit (the contraction, by 40 bits, of "binary digit") was coined by statistician John W. Tukey shortly after he joined von Neumann's project in November of 1945. The existence of a fundamental unit of communicable information, representing a single distinction between two alternatives, was defined rigorously by information theorist Claude Shannon in his then-secret Mathematical Theory of Cryptography of 1945, expanded into his Mathematical Theory of Communication of 1948. "Any difference that makes a difference" is how cybernetician Gregory Bateson translated Shannon's definition into informal terms. To a digital computer, the only difference that makes a difference is the difference between a zero and a one.

That two symbols were sufficient for encoding all communication had been established by Francis Bacon in 1623. "The transposition of two Letters by five placeings will be sufficient for 32 Differences [and] by this Art a way is opened, whereby a man may expresse and signifie the intentions of his minde, at any distance of place, by objects ... capable of a twofold difference onely," he wrote, before giving examples of how such binary coding could be conveyed at the speed of paper, the speed of sound, or the speed of light.

That zero and one were sufficient for logic as well as arithmetic was established by Gottfried Wilhelm Leibniz in 1679, following the lead given by Thomas Hobbes in his Computation, or Logique of 1656. "By Ratiocination, I mean computation," Hobbes had announced. "Now to compute, is either to collect the sum of many things that are added together, or to know what remains when one thing is taken out of another. Ratiocination, therefore is the same with Addition or Substraction; and if any man adde Multiplication and Division, I will not be against it, seeing ... that all Ratiocination is comprehended in these two operations of the minde." The new computer, for all its powers, was nothing more than a very fast adding machine, with a memory of 40,960 bits.

In March of 1953 there were 53 kilobytes of high-speed random-access memory on planet Earth. Five kilobytes were at the end of Olden Lane, 32 kilobytes were divided among the eight completed clones of the Institute for Advanced Study's computer, and 16 kilobytes were unevenly distributed across a half dozen other machines. Data, and the few rudimentary programs that existed, were exchanged at the speed of punched cards and paper tape. Each island in the new archipelago constituted a universe unto itself.

In 1936, logician Alan Turing had formalized the powers (and limitations) of digital computers by giving a precise description of a class of devices (including an obedient human being) that could read, write, remember, and erase marks on an unbounded supply of tape. These "Turing machines" were able to translate, in both directions, between bits embodied as structure (in space) and bits encoded as sequences (in time). Turing then demonstrated the existence of a Universal Computing Machine that, given sufficient time, sufficient tape, and a precise description, could emulate the behavior of any other computing machine. The results are independent of whether the instructions are executed by tennis balls or electrons, and whether the memory is stored in semiconductors or on paper tape. "Being digital should be of more interest than being electronic," Turing pointed out.

Von Neumann set out to build a Universal Turing Machine that would operate at electronic speeds. At its core was a 32-by-32-by-40-bit matrix of high-speed random-access memory -- the nucleus of all things digital ever since. "Random access" meant that all individual memory locations -- collectively constituting the machine's internal "state of mind" -- were equally accessible at any time. "High speed" meant that the memory was accessible at the speed of light, not the speed of sound. It was the removal of this constraint that unleashed the powers of Turing's otherwise impractical Universal Machine.

Electronic components were widely available in 1945, but digital behavior was the exception to the rule. Images were televised by scanning them into lines, not breaking them into bits. Radar delivered an analog display of echoes returned by the continuous sweep of a microwave beam. Hi-fi systems filled postwar living rooms with the warmth of analog recordings pressed into vinyl without any losses to digital approximation being introduced. Digital technologies -- Teletype, Morse code, punched card accounting machines -- were perceived as antiquated, low-fidelity, and slow. Analog ruled the world.

Excerpt from:
1953: The Year That Revolutionized Life, Death, and the Digital Bit