'Walking is great!' Heart-warming video shows brave six-year-old's first steps after 'miracle recovery' from tragic car crash

A brave six-year-old Gold Coast girl is being called a medical miracle after taking her first steps since a tragic car crash which killed her grandmother and left her 99.8 per cent brain dead eight months ago.

The crash happened in Port Macquarie, NSW and her mother, who was also injured in the crash, was told that her daughter Mackinlee Anderson might not survive her injuries. 

Doctors believed if she did survive the six-year-old may never be able to talk or walk again.

"Mackinlee had horrific injuries. She had a broken pelvis in three different places, she had a ruptured bladder, a dislocated hip and a broken femur," Ms Anderson told Nine News.

"That was all on the bottom half. And then she had an extreme brain injury, with extreme brain trauma.

"We were pretty much in a place there where the doctors thought there was no hope for her."

However, inspirational footage shows Mackinlee taking her first steps since the accident and talking as she does.

"Walking is the best, walking is great," the young girl says as she makes her way unassisted down a hallway.

Her proud mum is still amazed at the miracle recovery Mackinlee has made.

"Mackinlee amazes me every day with her determination. Every day the doctors say she can't do something and the next day she can do it.

"And she walks everywhere now, around the house, everywhere. You just can't keep her still."


Eruption of Hawaii's Kilauea Volcano imminent after warning level raised to red

A major eruption of Hawaii's Kilauea volcano is imminent according to the United States Geological Survey who've issued a red warning.

The USGS says the warning was changed from orange to red due to increased ash emission from Kilauea Volcano summit. 

Earlier this month the volcano began erupting when a number of fissures opened up. 

Resident of Leilani Estates and Lanipuna Gardens have already had to evacuate.

Ashfall and vog (volcanic air pollution) has been reported in Pahala, USGS said in a statement. 

"At any time, activity may become more explosive, increasing the intensity of ash production and producing ballistic projectiles near the vent."

ONN 1 News at 6 promo image
For more on this story, watch 1 NEWS at 6pm. Source: 1 NEWS


Facebook removing sexism and hate speech posts - but finding it difficult

Getting rid of racist, sexist and other hateful remarks on Facebook is more challenging than weeding out other types of unacceptable posts because computer programs still stumble over the nuances of human language, the company revealed overnight NZT.

Facebook also released statistics that quantified how pervasive fake accounts have become on its influential service, despite a long-standing policy requiring people to set up accounts under their real-life identities.

From October to December alone, Facebook disabled nearly 1.3 billion accounts - and that doesn't even count all the times the company blocked bogus profiles before they could be set up.

Had the company not shut down all those fake accounts, its audience of monthly users would have swelled beyond its current 2.2 billion and probably created more potentially offensive material for Facebook to weed out.

Facebook's self-assessment showed its screening system is far better at scrubbing graphic violence, gratuitous nudity and terrorist propaganda. Automated tools detected 86 percent to 99.5 percent of the violations Facebook identified in those categories.

For hate speech, Facebook's human reviewers and computer algorithms identified just 38 percent of the violations. The rest came after Facebook users flagged the offending content for review.

All told, Facebook took action on nearly 1.6 billion pieces of content during the six months ending in March, a tiny fraction of all the activity on its social network, according to the company.

The report marked Facebook's first breakdown on how much material it removes for violating its policies. It didn't disclose how long it takes Facebook to remove material violating its standards. The report also doesn't cover how much inappropriate content Facebook missed.

"Even if they remove 100 million posts that are offensive, there will be one or two that have some really bad stuff and those will be the ones everyone winds up talking about on the cable-TV news," said Timothy Carone, who teaches about technology at the University of Notre Dame.

The report also doesn't address how Facebook is tacking another vexing issue - the proliferation of fake news stories planted by Russian agents and other fabricators trying to sway elections and public opinion.

Fake accounts on Facebook have been drawing more attention because Russian agents used them to buy ads to try to influence the 2016 election in the U.S.

Even though it has been focusing on shutting down bogus accounts, Facebook has said that 3 to 4 percent of its active monthly users are fake. That means as many as 88 million fake Facebook accounts were still slipping through the cracks in the company's policing system through March.

It's not surprising that Facebook's automated programs have the greatest difficulty trying to figure out differences between permissible opinions and despicable language that crosses the line, Carone said.

"It's like trying to figure out the equivalent between screaming 'Fire!' in a crowded theater when there is none and the equivalent of saying something that is uncomfortable but qualifies as free speech," he said.

Facebook said it removed 2.5 million pieces of content deemed unacceptable hate speech during the first three months of this year, up from 1.6 million during the previous quarter. The company credited better detection, even as it said computer programs have trouble understanding context and tone of language.

Facebook took down 3.4 million pieces of graphic violence during the first three months of this year, nearly triple the 1.2 million during the previous three months. In this case, better detection was only part of the reason. Facebook said users were more aggressively posting images of violence in places like war-torn Syria.

The increased transparency comes as the Menlo Park, California, company tries to make amends for a privacy scandal triggered by loose policies that allowed a data-mining company with ties to President Donald Trump's 2016 campaign to harvest personal information on as many as 87 million users.

The content screening has nothing to do with privacy protection, though, and is aimed at maintaining a family-friendly atmosphere for users and advertisers.