Blog

  • Nokia wants you to make an Ovi app out of your Twitter feed, launches Ovi App Wizard

    To anyone who’s ever written a line of code, pre-canned/auto-generated apps are the scourge of the App Store world. For anyone who just wants to be able to say “Me? There’s an app for that! LOL!” without having to dedicate their lives to waging war against semi-colons and pouring through documentation, however, they’re a god send.

    Love’em or hate’em, Nokia has just embraced’em. They’ve just launched a new tool, Ovi App Wizard, purposed with turning any feed (be it RSS, Twitter, Facebook, or what have you) into an Ovi Store App in around 10 minutes.

    The process seems pretty simple: Name it, pick a few RSS feeds, choose the colors, and then set up the distribution options. If you’re looking to cash in, you’ve got two choices: toss in a pre-fabbed ad system and take a cut of the revenue, or — if you’re feeling ambitious about your work — you can pick a price. Hit the publish button, and about 24 hours later, blammo – you’re famous. Well, not famous. But you have an Ovi Store App. That’s almost the same thing, right?

    It’s always a good thing when an App Store operator makes their outlet more accessible — but this really just seems like a ploy to boost their number of apps through quantity-over-quality. If 10,000 people make an application out of their Twitter feed, that’s 10,000 new apps Nokia can boast about – but how many of those will actually see more than a handful of downloads? Is an app really an app if the only person who downloads it is your Mom?

    There’s a reason Apple is banning these sorts of apps.

    Check out Ovi App Wizard here.


  • I Want to Get My Kids to the Egg Child Care Center [Architecture]

    This is Discovery, a children nursery designed by German studio GPAC. Thirteen eggs scattered under a central glass canopy. I want to get my child there. It will match my Star Trek suit. Or my full body latex chicken outfit. More »







  • The Gulf of Mexico oil spill by the numbers

    oil spill satellite view

    (Photo: NASA)

    As officials continue to carry out clean-up efforts from the deadly April 20, 2010, BP oil spill, one thing is clear: Last week’s Gulf Coast spill is one of the worst in history.

    It’s tough to say just how much oil might leak out before the three major underwater leaks are stopped. Right now around 210,000 gallons of oil are pouring into the Gulf each day. That’s nowhere near the 11 million gallons of oil leaked into Alaska’s Prince William Sound by the Exxon Valdez tanker. Yet.

    But if the leaks aren’t repaired soon, the BP spill may overtake the iconic Valdez tragedy in volume. If oil continues to leak at its current rate, the BP spill will be larger than the Exxon spill by the third week in June.

    This isn’t the first time British Petroleum has run into trouble. In 2005, a massive explosion occurred at the company’s Texas City, Texas, refinery, killing 15 workers and injuring 170 others. The company was fined $87 million for negligence. A year later, the company ran into problems again when it was cited for leaking around 4,800 barrels of oil into Alaska’s Prudhoe Bay, due to a corroded stretch of pipeline. The company was warned about the corroded pipe four years earlier but did nothing to fix it — and was fined $20 million for ignoring opportunities to prevent the spill.

    In order to give you a better sense of just how big the unfolding tragedy in the Gulf is, we dug out some surprising statistics:

    11: Number of workers missing and presumed dead following the BP rig explosion. (Source: Huffington Post)

    5,000 barrels a day: Rate at which oil is leaking from the Deepwater Horizon rig — five times faster than was originally estimated. (Source: New York Times)

    $300 million: Estimated cost to BP to plug up the leaking oil spill, not including environmental clean-up costs. (Source: New York Times)

    $25 billion: Loss of market value to BP stock since last week’s rig explosion. (Source: Huffington Post).

    16: Miles off the coast of Louisiana the oil slick has crept. (Source: New York Times)

    At least 30: Species of birds the Audubon Society says are potentially threatened by the oil spill. These include marsh birds, ocean-dwelling birds, and migratory songbirds. All reside in “Important Bird Areas,” according to Audubon, designated because of their “essential habitat value.”

    Among the most vulnerable species is the brown pelican — the state bird of Louisiana — which was only recently removed from the endangered species list. The spill is especially devastating for bird populations because it coincides with the beginning of breeding season. (Source: Audubon Society)

    25 million: Number of birds that traverse the Gulf Coast per day and which are potentially at risk from the oil spill. According to the LA Times Greenspace blog, “Late spring is the peak time for neo-tropical songbirds moving from the Yucatan Peninsula to make their first landfall in Louisiana,” and “more than 70% of the country’s waterfowl frequent the gulf’s waters.”(Source: LA Times Greenspace blog)

    11 million: Number of gallons of oil leaked into Alaska’s Prince William Sound in 1989 by the Exxon Valdez oil leak. It is widely considered the worst oil spill in U.S. history, although a number of larger spills have happened around the world, including the 2002 Prestige spill off Spain. (Source: CBS News)

    400: Number of wildlife species threatened by the spill. Threatened species include sea life such as whales, tuna, and shrimp; dozens of species of birds; land animals such as the gray fox and white-tailed deer; and amphibians such as the alligator and the snapping turtle. (Source: New Orleans Times-Picayune)

    600 square miles: Latest reported size of the oil slick. In response to reports of the blooming size of the spill, Louisiana Governor Bobby Jindal declared a state of emergency, anticipating that it would reach Louisiana’s shores soon. (Source: CNN)

    $1.5 billion: Amount in insurance claims experts believe the BP spill will cost insurers. (Source: Business Week)

    More from The Daily Green

    Reprinted with permission of Hearst Communications, Inc.

    Check out Yahoo! Green on Twitter and Facebook.

  • Lawmakers Eye Race To The Top Funding; Senate passes School Reform Bill And Sends It To The House

    Education reporter Grace Merritt is following a school reform bill, which was debated by the state Senate today. Her story is below.

    HARTFORD — The state Senate approved a wide-ranging school reform bill Friday that would require high school students to take more math and science courses and, for the first time, foreign language classes to graduate.

    The bill, which would also link teacher evaluations to student performance and lift enrollment limits on charter schools, is designed to better position the state to compete for millions in federal Race to the Top school reform funding.

    Legislators and state education leaders have been working for weeks with teacher groups, charter school advocates, superintendents and others in “grueling” sessions to negotiate the bill, said state Sen. Tom Gaffey, D-Meriden, co-chairman of the education committee.

    While the bill was approved in the Senate, it’s expected to face a bumpy ride in the House. The Republican caucus is now reviewing the bill, said state Rep. Andrew Fleischmann, D-West Hartford, who co-sponsored the bill with Gaffey.

    The bill establishes a more rigorous high school curriculum designed to better prepare students for college and compete in a global economy.

    The new standards means students would have to earn a minimum of 25 credits to graduate, up from 20 now, including two language credits and one more credit each in math and science.

    Students would also have to take end-of-year exams for core courses to ensure that they’ve learned the material. Seniors will be required to complete a multidisciplinary “capstone project” — spelling the end of planned coasting for the second semester of senior year.

    The new requirements wouldn’t take effect until the class of 2018, a concession made to make the bill more palatable to opponents who characterize it as another unfunded mandate in a poor economy.

    “We certainly are not going to be in this economic condition in for too much longer,” Gaffey said.

    Besides, he said, the bill is designed to help the state win millions in federal stimulus money from the Race to the Top competition, which officials hope might bring as much as $192 million to Connecticut.

    Having failed to win funds in the first round of the Race to the Top competition, the state is revising its application to try again June 1. This time around, the state is hoping that the bill’s passage will strengthen the application by demonstrating the state’s commitment to school reform.

    The bill would also establish a new framework for teacher evaluations that would use indicators of student academic growth in assessing teacher performance. Under the plan, the state Board of Education would work with an advisory board consisting of representatives from teacher unions, school boards, and state and local education leaders to develop the evaluations.

    “I believe that this is a very reasonable approach to teacher evaluation at the present time,” said John Yrchik, executive director of the Connecticut Education Association, the largest statewide teachers union.

    In addition, the bill would also:

    – Life caps on student enrollment at high-performing charter schools.

    – Create an alternate route to certification for school administrators.

    – Allow struggling school districts to convert an existing school or build a new one as an “innovation school” to improve school performance.

  • CHART OF THE DAY: Now The Oil Slick Is A $33 Billion Disaster (BP, RIG, HAL, DRQ, TTI)

    The Deepwater Horizon explosion has been a nightmare for regulators, energy companies, and cleanup crews this week. As the situation continues to decline, oil service companies are taking a beating.

    Our chart below shows companies involved or affected by the Deep Water incident and the amount of market cap that they’ve lost since April 22nd, when the incident was first reported in the media. Altogether the combined amount of market cap lost totals a whopping $33.2 billion.

    Bonus: Everything you need to know about the Deepwater oil spill >

    chart of the day, deepwater spill affecting the market

    Join the conversation about this story »

  • When America was post-colonial | Gene Expression

    Below I stated:

    …until the late 20th century the majority of the ancestry of the white population of the republic descended from those who were counted in the 1790 census.

    A commenter questioned the assertion. The commenter was right to question it. My source was a 1992 paper that estimated that only in 1990 did the proportion of American ancestry which derived from those who arrived after the 1790 census exceeding 50%. In other words, if you ran the ancestors of all Americans back to 1790, a majority of that set would have been counted in the 1790 census (so people of mixed ancestry would contribute to the two components are weighted by their ancestry).

    The major issue here is that there is a difference between whites, and non-whites, especially before mass Asian and Latin American immigration post-1965, when white vs. non-white ~ white vs. black. Almost all the ancestors of black Americans who were black were already resident in the United States in 1790. A few years ago I read up on the history of American slavery and was surprised how genuinely indigenous the black American, slave and free, population was by the late 18th century (English speaking and Christian). There was an obvious reason why Southern slave-holders went along with the ban on importation of slaves which was due to kick in in the early decades of the republic: American blacks, unlike slave populations elsewhere in the New World, had endogenous natural increase. This explains part of the relative paucity of African aspects in their culture in relation to the blacks of Haiti or Brazil, where African-born individuals were still very substantial numerically at emancipation because of high attrition rates (it is sometimes asserted that the majority of blacks liberated during the Haitian Revolution were born in Africa. Likely a hyperbole, but it gets across the strength of connection).

    In any case, to estimate the white proportion attributable to 1790, I have to correct for the black proportion within the total. As an approximation I think it’s acceptable to simply attribute blacks as a whole to the proportion which had ancestors here in 1790 in full. I suspect a greater proportion of the black ancestry which post-dates 1790 would come from the white component of their heritage which simply isn’t of notice in American society for various reasons in any case (Henry Louis Gates Jr. is more white than he is black in terms of ancestry, but he’s the doyen of Africana Studies). So, assuming that blacks contribute to the 1790 and before component in full, I estimate that between 1910 and 1920 the majority of the ancestry of the white population shifted from 1790 and before, to after. Specifically, in 1910 51% of the ancestry could be traced to 1790 and before among whites, and in 1920 49%. In 1950 it was 47% 1790 and before. So I should have said early 20th century, not late. I wouldn’t be surprised though if the balance has started to shift in recent years, as many “white ethnic” groups (Jews, Italians, Irish, etc.) are more heavily concentrated in urban areas, while the most fertile white community in the United States, the Mormons of Utah, are also the most Old Stock Yankee in ancestry (I am aware that many Mormons are descended from European immigrants who converted in Europe and made the journey after conversion, but Mormons are still far more Old Stock Yankee than any group outside of interior New England).

  • New predictions for sea level rise

    Science Daily: Fossil coral data and temperature records derived from ice-core measurements have been used to place better constraints on future sea level rise, and to test sea level projections.

    The results are published in Nature Geoscience and predict that the amount of sea level rise by the end of this century will be between 7- 82 cm – depending on the amount of warming that occurs – a figure similar to that projected by the IPCC report of 2007.

    Placing limits on the amount of sea level rise over the next century is one of the most pressing challenges for climate scientists. The uncertainties around different methods to achieve accurate predictions are highly contentious because the response of the Greenland and Antarctic ice sheets to warming is not well understood.

    Dr Mark Siddall from the University of Bristol, together with colleagues from Switzerland and the US, used fossil coral data and temperature records derived from ice-core measurements to reconstruct sea level fluctuations in response to changing climate for the past 22,000 years, a period that covers the transition from glacial maximum to the warm Holocene interglacial period.

    By considering how sea level has responded to temperature since the end of the last glacial period, Siddall and colleagues predict that the amount of sea level rise by the end of this century will be similar to that projected by the Fourth Assessment Report of the Intergovernmental Panel on Climate Change (IPCC).

    Dr Siddall said: “Given that the two approaches are entirely independent of each other, this result strengthens the confidence with which one may interpret the IPCC results. It is of vital importance that this semi-empirical result, based on a wealth of data from fossil corals, converges so closely with the IPCC estimates.

    “Furthermore, as the time constant of the sea level response is 2,900 years, our model indicates that the impact of twentieth-century warming on sea level will continue for many centuries into the future. It will therefore constitute an important component of climate change in the future.”

    The IPCC used sophisticated climate models to carry out their analysis, whereas Siddall and colleagues used a simple, conceptual model which is trained to match the sea level changes that have occurred since the end of the last ice age.

    Read more>>

  • Bahn burner turned track rat: MTM Audi RS6 Clubsport

    Filed under: , , , , ,


    Audi RS6 Clubsport by MTM – Click above for image gallery

    Of all the products in the Audi arsenal, we’re not sure if the RS6 would be our track rat of choice. It’s twin-turbocharged V10, massive meats and grippy all-wheel drive system is surely the stuff of track-day dreams, but we’ve always been more enamored by the RS6’s brutal, road-going civility – it’s simply better suited to high-speed cross-country runs rather than full-on tarmac attacks. MTM appears to disagree.

    The German tuner of all things Audi has put the sedan variant under the knife to create the RS6 Clubsport, and as with everything from MTM, it’s a compellingly comprehensive package.

    The mods start with a new front lip, carbon fiber rear diffuser, an adjustable suspension and black 21-inch wheels shod in Michelin Pilot Sport PS2 tires. Inside, the factory thrones have been nixed in favor of a duo of carbon fiber Recaro buckets with six-point harness attached to a rear-mounted roll cage that’s sure to have displaced the rear seats.

    Engine upgrades are limited to a high-flow air filter and stainless steel exhaust, but with the help of MTM’s crack ECU tuners, they’ve been able to boost output of the turbo’d 5.0-liter V10 to 720 horsepower (at 6,360 rpm) and and 579 pound-feet of torque (between 2,750 and 6,280 rpm) – a sizable increase over the stock engine’s 571 hp and 479 lb-ft of twist. And if you option up the tires to one of the four available compounds (and why wouldn’t you?), a derestricted top speed of 211 mph can be yours. We think we’ll take ours in five-door form, sans the cage. Maybe in brown…

    [Source: MTM via LeBlogAuto]

    Bahn burner turned track rat: MTM Audi RS6 Clubsport originally appeared on Autoblog on Fri, 30 Apr 2010 15:31:00 EST. Please see our terms for use of feeds.

    Permalink | Email this | Comments

  • What’s More Accurate Than GPS? Photographs

    Locations identified within the 10 or 20 meters possible by GPS today are far too inaccurate — we need to know where we are we are right down to the millimeter! That was the gauntlet thrown down by Michael Liebhold, distinguished fellow at the Institute for the Future, speaking at a GigaOM Pro Bunker Session on location at the GigaOM office this week. With millimeter accuracy, augmented reality — digital information overlaid on a real-time view of the world — will actually become possible. “Right now we have all this toy AR,” said Liebhold. “This is useless.”

    So how do we get to millimeter accuracy? To find out, we followed up with Liebhold for a video interview. He said the most promising technique is to build model of the world using photographs, some of them geo-coded automatically, and the rest of them mapped using an understanding of where they are by comparing them to other images. So a photograph of vacationers in front of the Golden Gate bridge could be pinpointed in position using the precise angle of the orange arches in the background. Google Goggles is embarking on this very project, building a point cloud reference database using publicly available images like the ones from Flickr, said Liebhold, referencing remarks made by a member of the Goggles team at the recent Where 2.0 conference. (As is Microsoft, with its Photosynth product.)

    The Google project is scary, said Liebhold. Scary because of the privacy implications, I asked? No, he said, because if Google wants to do this, it will, and it will be hard to compete. Everyone wanting to use the most accurate location data will have to depend on Google.

    Liebhold did mention one promising startup effort in the space: Earthmine out of Berkeley, Calif., is building a set of street-view images captured in 3-D with every pixel geo-coded. (See our interview with them from a couple years back.)

    Intrigued as to how soon millimeter accuracy might happen and what it could enable? Here’s the video:

    Related content from GigaOM Pro (sub req’d):

    Report: Mobile Augmented Reality Today and Tomorrow

    Image courtesy of Flickr user jmlawlor

  • Cliq XT maintenance update being rolled out now

    Looks like the maintenance update that was being tested on the Cliq XT is now being rolled out to everybody. It’s in two parts, so you’ll need to do both of them. The update brings you fixes to:

    • Improves Overall Phone Performance
    • Visual Voice Mail Enhancements
    • Bluetooth Audio
    • Camera Correction
    • Screen/Display Performance
    • Text Messaging Improvements
    • Touch Screen Enhancement

    And we’re reminded that, no, it’s not Android 2.1. Gonna have to wait a little longer for that. [T-Mobile]

  • Climate debate gets ugly as world moves to curb CO2

    ABC News: Skeptics also accused the U.N.’s Intergovernmental Panel on Climate Change of supporting flawed science after several errors in a major 2007 report surfaced.

    The errors, including a reference to a non-peer reviewed study that Himalayan glaciers would melt by 2035, represent a fraction of the conclusions in the report, the main climate policy guide for governments, which is based on the work of thousands of scientists.

    The IPCC has defended its work and has ordered a review. Many governments, including the United States, Britain and Australia have also reiterated their faith in the IPCC.

    For climate scientists, truth and trust are at stake.

    “In general, the battle for public opinion is being lost,” said Kevin Trenberth, head of climate analysis at the National Center for Atmospheric Research (NCAR) in Boulder, Colorado. His emails were also hacked in the CRU incident.

    “There is so much mis-information and so many polarized attitudes that one can not even hold a rational discussion or debate. The facts are certainly lost or glossed over in many cases. The media have been a bust.”

    Schneider said the mainstream media had failed to do “its job of sorting out credible from non-credible and not giving all claimants of truth equal status at the bargaining table.”

    Across the Internet, the climate science debate is being played out in a myriad of climate skeptic sites and blogs as well as sites defending the science of human-induced climate change.

    One high-profile site is climatedepot.com, run by Marc Morano, a former aide to Republican Senator James Inhofe, who is an outspoken critic of climate change policies.

    Morano, who told Reuters he had also been the target of abusive emails, has been quoted as saying that climate scientists should be publicly flogged.

    “The global warming scientists need to feel and hear the public’s outrage at their shenanigans like “climategate” … There is no advocacy of violence or hint that people should threaten them,” Morano said, adding: “Public outrage is healthy.”

    Read more>>

  • Building Green and Healthy Places to Learn


    Almost one in five Americans are housed in schools for the better part of each day, but many of these schools offer toxic environments with poor daylight and are sited in far-off places, which means they are both unhealthy learning environments and contribute to sprawl (or unhealthy communities). Creating green and healthy schools which are in walkable, bikeable neighborhoods is key to increasing test scores and graduating children who can be future stewards of the environment. But how do we build green schools? This question and others were asked during the National Building Museum’s latest “For the Greener Good” discussion on sustainable schools.

    Joanne Silberner, Health Correspondent, National Public Radio, moderated a panel featuring Dr. Howard Frumkin, Director, National Center for Environmental Health / Agency for Toxic Substances and Disease Registry, U.S. Centers for Disease Control and Prevention; Steve Turckes, Director of K-12 Educational Facilities Group, Perkins+Will, and Glenn Cummings, Deputy Assistant Secretary, U.S. Department of Education.  

    Defining green schools

    According to Steve Turckes, Perkins + Will, sustainability was best defined by the Brundtland Report: it means meeting the need of the present without compromising the ability of future generations to meet their needs. While sustainable design isn’t unique to schools, green schools must have lower energy and water usage, daylighting, and up-to-date mechanical systems. To be sustainable, these schools must also be sited in walkable environments.

    Green schools should also teach sustainability and incorporate their green building features into the curriculum. “The school should be used as a tool to teach students about sustainability. A well-designed school can be an encyclopedia.”

    Glenn Cummings, Deputy Assistant Secretary, U.S. Department of Education, concurred that the physical elements of a building are important, but, more importantly, green schools can help students “make the connections between the impact of their personal choices and the environment.” Green schools can create environments conducive to learning while also serving as an instructional tool. “For instance, through maximizing solar orientation, you can teach ecology, meteorology, climatology – students can learn about science.”

    “Green + healthy is the real sweet spot,” said Dr. Howard Frumkin, U.S. Centers for Disease Control and Prevention. While many buildings are green, some may not be healthy — you want to have both. People need physical activity to be healthy. 

    Data on green schools

    Frumkin said you can look at the rates of teacher and student absenteeism, asthma or other sicknesses, as well as test performance metrics and disciplinary problems at green and non-green schools. There’s positive “outcome data” out there, which can also be reviewed along with cost data. “Hopefully, once schools see this data, going green will just seem like good business sense.”

    On cost, Turckes referred to Greg Katz’s data (see earlier post), which shows the cost premium of any green building at around 2 percent. “However, that data is now four years old. Right now, we think there is no extra cost to building green. With the extra energy savings and health benefits, the question is why wouldn’t you do it?”

    Obstacles preventing the growth of green schools

    It seems amazing, but according to Cummings, 14 million children go to school each day in “outright dangerous” schools. As you see in Washington, D.C. before the school term starts, schools scramble to “remedy buildings so they will be legal to occupy.” The U.S. has hundreds of thousands of school buildings, many of which were created up to 50 years ago. “The real challenge is retrofitting older buildings so they can be turned into green buildings.”

    Turckes added that the vast majority of those older buildings haven’t benefitted from newer technologies. “There have been huge advances in building technologies.” But still, many older buildings are mold-infested and feature outmoded air ventilation systems. On the positive side, he said it’s not impossible to revamp an older school. “We just figured out how to integrate an energy-efficient system into a vintage 1960′s school.” 

    The perception of the higher cost of green schools may present an obstacle. “There is no extra cost to sustainable design — it’s embedded, a form-driver. We use integrated interdisciplinary teams to do design up front, thereby removing inefficiences and saving costs.” However, with tight budgets, some schools still see commissioning agent costs and U.S. Green Building Council (USGBC) accreditation costs as a burden. To fight this: Turckes says some schools use the sustainable design process, but go green without getting the certificate.

    On top of these challenges, local governments sometimes make it difficult for themselves. “Budget structures can create obstacles to green buildings — maintenance and operations are often in two separate pots.”

    Addressing health problem in schools now

    If a school can’t afford to retrofit, Frumkin said, they can replace toxic cleaning supplies, ensure they are keeping HVAC maintenance up-to-date, and continually discarding art / science labs. “With budget cuts, sloppiness can set in.” To prevent this sloppiness, Frumkin recommended the E.P.A.’s school audit tools. “Greeing a school doesn’t have to cost a fortune. Safer cleaning materials and doormats so students can wipe their feet at the entrances of school buildings can help keep toxins out.”

    Frumkin also said educational workers — teachers, administrators, janitors — face enormous occupational hazards working in these older schools. On the positive side though, he said, there were some good studies looking at these workers and showing the positive effects of moving to green schools.”They are an easy population to study because they stay in buildings for a long time.” Another benefit: “Green schools also retain staff better.”

    Designing a healthy environment around schools

    In terms of student health, siting schools in walkable neighborhoods is key to fighting obesity, diabetes, heart disease. Sedentary lifestyles set a bad example. Frumkin added, “unfortunately, now a small minority of children walk or bike to school. This didn’t used to be the case.”

    To remedy this, schools basically need Smart Growth neighborhoods, with high levels of connectivity, inter-connected grid-like networks of streets, careful intersection engineering, roads with traffic calming measures, and dedicated lanes for children to walk to school. “Sidewalks and paths need to be safe for children.”

    Frumkin added changes are needed to both the built environment, policy, and behavior. He cited Safe Routes to Schools, and No Child Left Inside, two major movements, as very positive.  He also called for high fructose-drinks and vending machines to be removed from schools.

    The future of sustainability

    Turckes said all buildings need to be carbon neutral or net-zero in the near future. “This is where we need to get.” He cited the Architecture 2030 campaign. The U.S. in particular needs to get its act together. “We have 5 percent of the world’s population, but use 25 percent of its resources. This needs to be dramatically improved.”

    Cummings said sustainability curriculum needs to be further incorporated into schools so the U.S. can scale up for the green jobs of the future. (see earlier post). One audience member noted, however, that even if sustainability is on teachers’ radars, it’s often not included in “No Child Left Inside”-mandated tests, so it gets bumped.

    “The recently passed health reform package will lead to a new focus on the prevention of illness,” Frumkin said. “Green buildings, sustainable neighborhoods, healthy lifestyles — my dream is to marry all these together.” In the future, he said, perhaps health specialists will also be involved in green building design from the onset.

    Image credit: 2010 Professional Honor Award, General Design. Nueva School, Hillsborough, CA U.S.A. Andrea Cochran Landscape Architecture, San Francisco, U.S.A.

  • Scientists in Bangladesh question IPCC’s claims on flooding

    Greenwire: Scientists in Bangladesh have released a study indicating that higher sea levels would not cause devastating flooding in the low-lying country, as suggested by the Intergovernmental Panel on Climate Change’s most recent report.

    The U.N.-affiliated panel has taken pressure in recent months for mistakes in its 2007 report on global climate change. According to scientists at the Dhaka-based Center for Environment and Geographic Information Services, that report may have erred in its claim that a 3-foot rise in sea levels would flood 17 percent of Bangladesh and drive 20 million people from their homes by 2050.

    Himalayan rivers carry 1 billion tons of sediment into Bangladesh every year, the researchers say, shoring up coastlines against rising sea levels. Even if sea levels rise 3 feet, the accumulation of sediment should mitigate the problem, said Maminul Haque Sarker, the director of the environment center and the study’s lead researcher.

    “Sediments have been shaping Bangladesh’s coast for thousands of years,” he said. “Studies on the effects of climate change in Bangladesh, including those quoted by the IPCC, did not consider the role of sediment in the growth and adjustment process of the country’s coast and rivers to the sea level rise.”

    Rajendra Pachauri, chairman of the IPCC, defended the panel’s conclusions on flooding in Bangladesh. He also criticized broader efforts to discredit the report, saying that “one single error doesn’t take anything away from the major findings of the report.”

    Read more>>

  • Blog Review: Beaker – A Medical Research Blog

    The blog reviewed here is ‘Beaker – A Medical Research Blog‘ which I came across from one of Boraz’s Twitter posts.

    Appearance and Design

    The blog features a grey background with a diagonal white lined design. The central pane containing the text has a white background. The title pane features a photograph of two scientists focusing on a task next to the title of the blog. The articles are of a good length, informative, titled with author details, comment enabled , tagged and posted in specific categories. The articles are generously illustrated with high quality and relevant photographs including micrographs. There is one caveat however. In many of the articles, the reader must click on the ‘read the rest of the entry’ link at the bottom of the summary for each post in order to read the full post. The blog can be navigated using functions in the right hand panel – according to tags, categories and recent posts. There is also a search box. There are external links as well as a link to donate to the Sanford Burnham Institute.

    Content

    As would be expected from a large institution, there are frequent events (e.g visits by influential figures to the institute) and reports on research studies that are being undertaken. In this post for instance, the author succinctly describes an approach to designing new molecules that is being pursued at the institute by one researcher.  This post explains the role of messenger RNA and refers to a symposia on mRNA being held at the Institute.  As the blog is relatively young there are 15 posts at the time of writing.

    Conclusions

    This is a young blog which demonstrates how an institution can effectively represent itself through the use of social media. Indeed with a large number of employees and students, an Institution has a marked advantage if it is able to efficiently leverage its resources through the medium of social media. For the reader, the posts are well written and offer insights into the interesting research that is being undertaken at the Institute. This blog would particularly appeal to those interested in working or studying at the institute, readers with an interest in research or those interested in representing their institute through social media.

    Call for Authors: If you are interested in writing an article or series of articles for this blog please write to the e-mail address below. Copyright can be retained. Index: An index of the site can be found here. The page contains links to all of the articles in the blog in chronological order. Twitter: You can follow ‘The Amazing World of Psychiatry’ Twitter by clicking on this link. Podcast: You can listen to this post on Odiogo by clicking on this link (there may be a small delay between publishing of the blog article and the availability of the podcast). It is available for a limited period. TAWOP Channel: You can follow the TAWOP Channel on YouTube by clicking on this link. Responses: If you have any comments, you can leave them below or alternatively e-mail [email protected]. Disclaimer: The comments made here represent the opinions of the author and do not represent the profession or any body/organisation. The comments made here are not meant as a source of medical advice and those seeking medical advice are advised to consult with their own doctor. The author is not responsible for the contents of any external sites that are linked to in this blog.

  • Veoh’s Legal Fight Against Universal Continues… With Mystery Funder

    Last fall, you may recall, there was a very important — and very well argued — ruling against Universal Music in its attempt to sue video site Veoh. The court found that Veoh was clearly protected by the DMCA’s safe harbor provisions. The ruling was important on a few different points, especially since the entertainment industry has been working overtime to try to change the definition of the DMCA’s safe harbors to make them effectively meaningless. Thankfully, the court put a stop to that. However, things got complicated in February, when Veoh declared Chapter 7 bankruptcy. We wondered what would happen to the appeal that Universal Music was filing, and Eriq Gardner answered:


    If Veoh declares Chapter 7, a bankruptcy judge would issue an automatic stay in the case. UMG would likely file a motion with the bankruptcy court seeking relief from the stay to perfect its appeal. The trustee would engage legal counsel and make financial arrangements to cover the costs of defending the case before the 9th Circuit.

    However, that’s not quite what happened. The case is moving forward (with the same lawyers for Veoh — even though the company doesn’t exist), but the company never actually filed Chapter 7 bankruptcy. Joe Mullin has the latest details, which don’t clear up much. Instead of filing bankruptcy, at the last minute, it sold its assets to an Israeli company, Qlipso — but the lawsuit liabilities were separate. So, basically it’s a bit of a mystery who’s funding the ongoing lawsuit:


    On Thursday, Elkin confirmed to Corporate Counsel that he will represent the Veoh side on appeal, even though Veoh has ceased to exist as an operating company. Elkin said he is being paid to continue handling the case, which he says has consumed him for the past three years, but declined to comment on who is paying him. He said he and his team are “working mightily” to prepare their reply brief, which is due May 20.

    Just weeks before Veoh went out of business, I’d been told that the company was about to secure new funding solely to prop it up to fight this legal battle. So it’s interesting that there does appear to be funding, even if no one’s saying where it’s coming from. Of course, it wouldn’t be too hard to come up with a pretty short list of probable funders…

    That said, Mullin’s piece also goes through UMG’s appeals filing, and it’s a doozy. It effectively says that the DMCA’s safe harbors don’t exist, because Universal Music finds them inconvenient. I’m not kidding:


    [UMG] must incur the enormous expense of constantly monitoring Veoh’s internet site to identify infringing content and request its removal in order to protect their property. And the task is not limited to monitoring Veoh alone. Rather, it is geometrically larger since thousands of comparable websites must also be monitored. The task is ultimately Sisyphean; because Veoh’s site, like others’ is dynamic and changes day-to-day or hour-to-hour [and] as users upload more material, the task of identifying and sending notifications requesting the removal of copyrighted works would amount to an unending version of the children’s game of “Whack-A-Mole.”

    What Universal fails to point out is that if the process is hard for it, it’s actually infinitely harder for Veoh, and that’s because Veoh has no way of knowing for sure if content is infringing or not. As Mullin points out, Universal’s argument is effectively the same one that Tiffany has made over and over and over again against eBay — losing every time. It’s the argument that because it’s too inconvenient for rightsholders to police their rights, the courts should arbitrarily force service providers to do so — even as they have no insight into what’s really infringing and what’s not. In fact, you could argue that Tiffany had a stronger case, in that there aren’t safe harbors when it comes to trademark issues. Universal has a huge uphill battle here.

    Permalink | Comments | Email This Story





  • Dell Aero Has Google Apps After All

    Here’s some good news for you Android fans with AT&T service.  The Dell Aero will have the standard fare of Google applications and services.  This includes the Android Market,  YouTube, Google Maps, Google Talk, Gmail, and Google Search. It was initially reported that AT&T had taken steps to yank out of some the Google in favor of Yahoo, similar to the Backflip experience

    For those unfamiliar with the handset, the touch-only Aero features a 5 megapixel camera and a 3.5″ nHD display.  No official launch date is known as of yet.  Here’s to hoping AT&T lets users add apps from outside of the Android Market.

    Might We Suggest…

    • AT&T Planning to Lock Down all Android Handsets?
      As CTIA was kicking off last week, AT&T and Dell unveiled the forthcoming Aero handset. The sleek touch-only device which features a 3.5-inch screen and 5 megapixel camera may have something in co…


  • Scosche Takes A Crack at a Dashboard iPad Mount [Ipad]

    Now that the iPad 3G is on sale—complete with a GPS chip inside—you might want to mount it in your car. But nobody yet makes a good car solution. Scosche is gonna try—but there are issues. More »







  • Acura recalling 167,000 TSX sedans for fire risk

    2005 Acura TSX

    Acura announced that it is recalling 167,000 TSX sedans to replace a power steering hose, o-ring gasket and fluid. The affected models include 2004-2008 TSX units sold in the United States with the 2.4L inline 4-cylinder engine.

    “The recall is being conducted due to the potential for premature deterioration of the exterior surface of the hose as a result of prolonged exposure to high under-hood temperatures,” Acura said in a statement. “In the event of excessive deterioration, a crack in the hose could develop, which could allow power steering fluid to leak or spray on a hot exhaust component potentially generating smoke or a fire.”

    Acura said it has received only one incident resulting in a fire.

    Owners of affected vehicles will be informed by the end of May. Acura TSX owners can also go to owners.acura.com/recalls or may call (800) 382-2238 for more information.

    – By: Kap Shah


  • WHOA: Oil Spill May Be 5X Previous Estimates, And Could Pass Exxon Valdez TOMORROW

    oil spill birds

    The Deepwater Horizon oil leak could be leaking at a rate of 25,000 barrels a day, five times yesterday’s estimate. The leak rate is 25 times BP’s initial estimate of 1,000 barrels per day.

    If the new numbers are correct, the Gulf of Mexico spill is will pass Exxon Valdez tomorrow. The 1989 spill involved 260,000 barrels of oil.

    Get Up To Speed On The Oil Spill >

    Join the conversation about this story »