proba
float64 0.5
1
| text
stringlengths 16
182k
| timestamp
timestamp[s] | url
stringlengths 15
1.72k
|
---|---|---|---|
0.997928 |
Would a new Pinkston High really cost $130 million? Hard to say.
When Pinkston High School opened in 1964, Dallas ISD leaders said it marked the start of a "great tradition of education" in West Dallas.
Now, after decades of wear and tear, current district leaders say it's time to replace Pinkston with a school built for 21st-century academic needs.
The district's $1.6 billion bond proposal, which voters will decide Tuesday, includes $130 million for a new Pinkston. Supporters envision a modern campus to prepare students from across DISD for careers that could include architecture, health, automotive technology or engineering.
"There’s an excitement about making it a career-tech [school] with all the great things that are happening in West Dallas," Superintendent Michael Hinojosa said.
Skeptics question the cost and need for the new school. Pinkston’s enrollment has languished at around 1,000 students, just half of what the school can hold. It would be cheaper to fix the old building — about $42 million, according to an independent report from 2013.
Trustee Lew Blackburn, whose area includes Pinkston, said he supports the overall bond program, which would fund repairs, classroom additions and several new or replacement schools across DISD — and without requiring a tax-rate increase, district officials say. But Blackburn has questions about the new Pinkston’s price tag. It would be the district’s most expensive school yet.
“I’m not sure that we need $130 million for a high school,” Blackburn said. “I would rather we spend $80 million, maybe $100 million.” The remaining dollars could pay for repairs in other West Dallas schools, he said.
Hinojosa said he understands the concerns. He said he believes that the people who estimated the costs (district bond officials, working with a community advisory group) erred on the cautious side so they wouldn't come up short.
"They're trying to make sure that there's enough of a cushion," Hinojosa said, adding that he'd be surprised if a new Pinkston cost the full $130 million. Like Blackburn, Hinojosa said he imagines some of the savings going back into other West Dallas schools.
So how much should a high school cost? Experts say it’s hard to nail down a figure because of all the variables. How big will the school be? What high-tech bells and whistles will it have? When and where will it be built?
The Texas comptroller’s office tracks school construction costs, but only through 2013. The most expensive school on that list was Centennial High School in Burleson ISD, which opened in 2010 at a cost of $106 million. That’s roughly $120 million in today’s dollars.
Around North Texas, costs for future campuses vary. Denton ISD has budgeted $157 million for a high school that’s scheduled to open next year and would hold up to 2,400 students. Frisco ISD plans to spend $110 million on a high school that would open around 2019 and hold 2,100 students.
In Houston ISD, voters approved a $1.9 billion bond in 2012. Now district officials project they'll run more than $200 million short. Superintendent Terry Grier cites rising construction costs, while critics say HISD hasn't managed all the money wisely.
The $130 million estimate for Pinkston is based on a high school with 325,000 square feet and up to 2,100 students, said Ed Levine, DISD's executive director of construction services. It's also based on a 2018 construction date, which would be halfway through the bond program. The total includes non-construction costs like permits and architecture fees, but not the cost to acquire land.
“There’s actually a very objective formula that they went through,” said Craig Reynolds, a Dallas architect who co-chairs an advisory task force that reviewed DISD's future building needs.
| 2019-04-24T02:19:55 |
https://www.dallasnews.com/news/news/2015/10/29/would-a-new-pinkston-high-really-cost-130-million-hard-to-say
|
0.999999 |
What is the difference between a router and an access point/hub?
at work, and I have been told that the company would like BT-broadband (to be installed shortly) to be shared. My access point does NOT contain a router, however I understand that I need a router to share this connection. Is this correct? Which one would you recommend for me to share a BT-Broadband connection? Most of the ones that I have seen on the D-Link Web site are access-points and routers combined but I don't think I need an access-point because I've already got one. A wireless AP bridges traffic from an 802.11 wireless LAN onto an 802.3 Ethernet LAN. The AP lets several 802.11 stations connect to your wired LAN, similar to the way that a hub or switch lets Ethernet stations connect.
A router relays network traffic (IP packets) between two or more subnets. A subnet can be composed of one or more LAN segments. All hosts in a subnet are assigned IP addresses from the same range, which allows traffic to be routed to and from the right subnet. Once the traffic hits the destination subnet, LAN frame forwarding delivers the traffic to its final destination.
Many vendors (D-Link, Linksys, Netgear, etc) sell both wireless APs and wireless routers. A wireless router is just a wireless AP with a BUILT-IN router. Instead of buying two separate devices, you can buy one device that does both. If you only need one AP and one router, this is usually cheaper and less difficult to set up. That's why wireless routers are popular with home users that have small networks.
Add a dual-Ethernet broadband gateway behind your modem like the LinkSys BEFSR11 EtherFast Cable/DSL Router. Gateways like these are inexpensive and provide basic broadband connection sharing.
Add a small business firewall behind your modem like the SonicWALL SOHO3. SMB firewalls provide higher throughput connection sharing plus more robust firewall and VPN features.
| 2019-04-22T19:57:11 |
https://searchnetworking.techtarget.com/answer/What-is-the-difference-between-a-router-and-an-access-point-hub
|
0.999999 |
Text reads from left to right: Whoa some people in other countries are comparing Bush to Adolph Hitler because of his warmongering.
That's preposterous. Even I would never compare Bush to Hitler.
I mean Hitler was democratically elected wasn't he?
| 2019-04-22T01:19:58 |
http://www.topplebush.com/cartoons46.shtml
|
0.997507 |
Lauren: What a question! So, I'm fascinated by this impending marriage between Meghan Markle and Prince Harry. Yes, because she is bi-racial but also because she's American. It's a standard American fairytale, of which I am often dismissive, but this idea of Harry spending time with Meghan's African-American mother and perhaps one day fathering little afro-wearing redheads both tickles me and frightens the hell out of me.
After the engagement was announced, there was some talk of Queen Charlotte, wife of George III, being the real first Black member of the British royal family. Queen Charlotte was considered very unfair, "ugly" is the word some of her contemporaries used and there seemed to be more talk than usual about this. I've traveled all over Europe and have been on those horribly long old palace visits and I have seen some "ugly" people in those paintings! And when I put a painting of Queen Charlotte next to say Anne of Austria or Marguerite de Valois of France, I don't see any of them more fair or more "ugly" than the other.
So, this has me convinced that perhaps some of those historians suggesting that she is a descendant of a Moor may be, in fact, correct. Because this is what Europeans were apt to do with a woman who had negroid features--call her ugly. What was her life like as talk swirled around her of her ugliness and her blackness? King George III was known as "Mad King George" so Queen Charlotte, while looked down upon by the court, was also dealing with this husband who by all accounts she loved very much and his declining mental health and the thirteen or so children they had together! Perhaps there's no story there, but if I could have Hilary Mantel's gift for researching, her dogged pursuit of historical data, added to my sensibilities as a Black woman and somehow meld both our approaches to storytelling, we might have one heck of a novel!
Greer: Yes! I would read that novel every day of the week and twice on Sundays. There's got to be a story in Charlotte's perspective. Next question: How would you describe what you write?
Lauren: I have this nagging itch to write women back into history. The New York Times recently began this series where they go back to look at all the obituaries of women who were overlooked by them. And I stress "overlooked by them" because I believe many of those women were and still are celebrated despite being ignored by the Times. With that said, this series reminded me of all the women of color in history who people don't even know to celebrate. Women who weren't forgotten as much as never seen. My grandmother was one of those women. She came to this country, fleeing an abusive husband, leaving her children behind in Trinidad, and within a year, making only $50/wk, managed to save enough to pay airfare for two of the six and then the year after for two more. This was remarkable for an uneducated woman from the rural countryside of Trinidad! And yet this story hadn't been told until I wrote about her in my first novel 'Til the Well Runs Dry. She didn't win any awards in her life, she scrubbed toilets in a New York hospital until she retired, and yet, she is my definition of resilience and fortitude. I write about ordinary women who find ways to joy and hope even while living under remarkably difficult circumstances.
Greer: Fantastic. What's your next book about and when will we see it?
Lauren: My next project, currently titled One True Place is set between colonial Trinidad and what was then known as the Louisiana Territory in North America. It spans nearly thirty-five years and tells the story of a proud and dignified family whose journey is upended by the arrival of both the English settlers and a gold-scavenging stranger. I love the idea that I've written my own kind of Western. I absolutely adore this book and I think my readers will too. It has all the elements I look for in a good book--community intrigue, a strong-willed woman at its core, it takes readers to a place they haven't been, there's love and betrayal, an unrelenting familial bond and of course, good history! Grove/Atlantic is publishing it and with any luck, they will have it in stores by Fall 2019.
Lauren: Aah..question for Greer--When you're riddled with self-doubt and maybe you feel like your book didn't do as well as you'd hoped or the reviews weren't as good or maybe your agent and editor don't love a part of something you've written as much as you, to what or to whom do you turn?
Greer: First of all, I love that your question isn't "if" I'm riddled with self-doubt, but "when." Because aren't we all like that? When I'm writing a new book, I am completely convinced one day that it's genius, and the next day I'm equally convinced it's trash. So I definitely have those self-doubt moments, and those moments of "Why is this person getting a review in the Times and I'm not? Why is someone else's book I didn't personally love getting tons of attention? Why doesn't my agent think my book is ready to send out on submission yet when I am just so freaking tired of rewriting it?" Basically every stage of the process has infinite potential for self-doubt. Yay, publishing!
And it's honestly my fellow authors I turn to. Sometimes explicitly, to ask a question about how to go forward, to ask advice: "Have you ever been in this situation? What did you do?" And sometimes just reaching out to them, celebrating the good stuff when it happens -- theirs or mine! -- gives me the confidence and strength to move forward when the stuff is not so good. Some would call it a network, but I feel like that makes it sound a little mercenary, when it's not. It's a community. None of this would be worth doing if it weren't for that community.
For more about Lauren and her books, visit http://www.laurenfrancissharma.com.
| 2019-04-21T05:00:50 |
http://www.greermacallister.com/blog/2018/3/28/womenshistoryreads-interview-lauren-francis-sharma
|
0.999998 |
title : NFL Rumors: Philadelphia Eagles Sam Bradford Refuses To Mentor Carson Wentz; Tim Tebow to The Rescue?
summary : The Philadelphia Eagles had quite a shake up in their team this NFL offseason. The football squad saw the departure of two key figures who were initially expected to bring them to the Super Bowl - Chip Kelly, who is now with the San Francisco 49ers, and DeMarco Murray, who signed with the Tennessee Titans recently. They also lost Mark Sanchez who currently attempts to follow the footsteps of the great Peyton Manning in the Denver Broncos. It appears, however, that the time for losing members of the Philadelphia Eagles family is not yet over.
summary : The NFL draft 2016 season is an exciting time not only for Heisman Trophy nominees but also for football stars who wish to make it to the big league. With the Dallas Cowboys, San Francisco 49ers, Philadelphia Eagles, Carolina Panthers and New York Jets actively seeking fresh talents in their roster, the likes of Carson Wentz, Joey Bosa and Paxton Lynch will not go unnoticed. In fact, even Super Bowl 50 champs Denver Broncos are looking into ways to add players worthy to be the successor of Peyton Manning.
summary : The blossoming romance between the Dallas Cowboys and Joey Bosa might come into fruition soon as the NFL draft season starts soon. Although the Ohio State star has been plagued by attitude concerns in the past, it looks like the Tony Romo team chooses to overlook his character problems and focus solely on his talent. That is, Jerry Jones appears to be applying the same principle he used when he signed Greg Hardy and caused him to be linked to Johnny Manziel - talent above everything else.
summary : While the San Francisco 49ers and Colin Kaepernick continue their negotiations, reports are circulating that Chip Kelly might just have found the perfect solution for the dilemma - kick out Kaepernick from the roster and sign a fresh star in the form of Paxton Lynch.
summary : The third week of the 2015 NFL preseason games will see the San Francisco 49ers facing off with the Denver Broncos on Saturday (August 29) at 1:00 pm (MHT). The Week 3 preseason match of the two NFL teams will feature the Denver Broncos playing host to the San Francisco 49ers at the Sports Authority Field at Mile High in Denver. Should the performances of the Denver Broncos and the San Francisco 49ers be used as basis to predict which NFL team would go home victories in this 2015 preseason match, then the Peyton Manning-led team are contenders for the prize.
summary : The Minnesota Vikings are rounding up the new players the NFL team acquired during the offseason and it looks like free agent Stephen Goodin is the latest star to join the Adrian Peterson-led group. In a press release on Tuesday, the Minnesota Vikings revealed that offensive tackle Phil Loadholt was transferred to the injured reserve list following the Achilles heel injury the NFL star incurred during the game against the Tampa Bay Buccahneers over the weekend. The Saturday NFL pre-season match saw the Minnesota Viking triumph over the Tampa Bay Buccahneers .
summary : Amid the efforts of the Pittsburgh Steelers to firm up the NFL team's opening roster comes yet another injury that posed a step backwards for the group that has been gunning to take home the Super Bowl title since 2008. The weekend saw an injury suffered by Pittsburgh Steelers veteran kicker Shaun Suisham during a match against the Minnesota Vikings. According to NFL reporter Conor Orr, the Pittsburgh Steelers kicker suffered from a torn anterior cruciate ligament (ACL) in his left knee. Due to this injury, Suisham is reported to be out for the entire 2015 NFL season.
summary : The Baltimore Ravens have cancelled their NFL draft party, which was supposed to take place on Thursday, due to the intense riots and protests in Baltimore surrounding Freddie Gray, who died in police custody on April 19. In addition, Ravens legend Ray Lewis and basketball player Carmelo Anthony have pleaded to Baltimore residents to halt the social unrest in their city.
summary : In the aftermath of Oregon's defeat to Ohio State 42-20 on Monday for the college football championship, quarterback Marcus Mariota has decided to end his collegiate run and pursue a career in the NFL.
| 2019-04-19T10:14:24 |
https://www.gospelherald.com/archives/articles/tags/nfl-draft
|
0.999998 |
a. Classify the interests: Aloysius has a fee simple determinable (AKA a fee simple subject to an executory limitation) and Bill has a shifting executory interest.
b. Apply the Rules: Bill’s executory interest will vest at the latest upon Aloysius’s death, so there is no Rule Against Perpetuities problem.
a. Classify the interests: Agatha has a life estate. Barbara has a contingent remainder in fee simple absolute. Oswald has a reversion in fee simple absolute.
b. Apply the Rules: We will know whether Agatha has died unmarried by the time of her death, so there is no Rule Against Perpetuities problem here.
a. Classify the interests: Aman has a life estate, Community Hospital has a contingent remainder in fee simple determinable, and O’Neil retains the possibility of reverter.
b. Apply the Rules: Under the common law Rule Against Perpetuities, there is actually no problem when the gift is phrased this way. The remainder in the hospital will become possessory at Aman’s death. Also, the Rule Against Perpetuities doesn’t destroy reversionary interests like O’Neil’s possibility of reverter.
a. Classify the interests: Anthony has a life estate. Blake has a fee simple subject to a condition subsequent (AKA fee simple subject to an executory limitation). Chelsea has a shifting executory interest in fee simple absolute.
a. Classify the interests: Agnes has a fee simple determinable (AKA fee simple subject to an executory limitation), Brad has a shifting executory interest for life, and Oswald has a reversion in fee simple absolute and the possibility of reverter (or maybe just the latter).
b. Apply the Rules: Though it may be hundreds of years before liquor is served on the premises, Brad’s executory interest will either vest or fail upon his death, so there is no Rule Against Perpetuities problem.
a. Classify the interests: Alana has a fee simple determinable (AKA fee simple subject to an executory limitation), and Barclay has a shifting executory interest in fee simple absolute.
c. Reclassify the interests: Alana has a fee simple determinable and Oswald has the possibility of reverter.
d. Under Illinois law: There is no effect and the executory interest to Barclay still fails.
a. Classify the interests: Annunzio has a fee simple subject to a condition subsequent (AKA fee simple subject to an executory limitation) and Boaz has a shifting executory interest in fee simple absolute.
b. Apply the Rules: There is no Rule Against Perpetuities problem because the executory interest will vest or fail by Annunzio’s death.
a. Classify the interests: Otero has a fee simple subject to an executory limitation (AKA fee simple subject to a springing executory interest) and Aleisha has a springing executory interest in fee simple absolute.
b. Apply the Rules: There is no Rule Against Perpetuities problem here because Aleisha will either die or turn 21 within 21 years.
a. Classify the interests: Otero’s widow has a life estate. Otero’s children alive at the time of his widow’s death have a contingent remainder in fee simple absolute. Otero’s heirs have a reversion in fee simple absolute.
a. Classify the interests: Arturo has a life estate, SLU has a vested remainder subject to total divestment (AKA vested remainder in fee simple determinable), and the APA has either the possibility of reverter if it is devisable, or nothing if it isn’t. Otero’s heirs get the possibility of reverter if it can’t be devised.
b. Apply the Rules: The gift to SLU is fine because it will vest at Arturo’s death. Otero retained the possibility of reverter in the first sentence and then gave it to APA in the second sentence and he’s allowed to do so, then the possibility of reverter will survive because the Rule Against Perpetuities doesn’t kill reversionary interests.
c. Reclassify the interests: If the executory interest fails, Arturo has a life estate, SLU has a vested remainder subject to divestment, and Otero has the possibility of reverter.
d. Under Illinois law: There’s no difference here.
b. Apply the Rules: This isn’t quite a Rule in Shelley’s Case situation. Abner’s heirs are not named as remaindermen. Abner’s children are so named. So we can’t collapse the two gifts into a fee simple absolute for Abner.
a. Classify the interests: The Salsich Museum has a fee simple determinable in the St. Louis property. Otero has the possibility of reverter. If the possibility of reverter is transferable by deed and we interpret the gift to include future interests, then the Art Institute gets the possibility of reverter as well as all of Otero’s real property outside of St. Louis in fee simple absolute.
b. Apply the Rules: A possibility of reverter never fails under the Rule Against Perpetuities. There are no contingent interests in this example.
a. Classify the interests: It sounds like Antonio has a life estate even though the language is that of a fee simple determinable to some degree. But we’ll call it a life estate. Then his wife has a contingent remainder for life. Finally, the children have either a contingent remainder or a vested remainder subject to open in fee simple absolute (depending on if there are any children around now).
b. Apply the Rules: Antonio and his wife will be ascertained upon Ochs’s death. Antonio’s children will be ascertained upon Antontio’s death. Therefore, there is no problem with any of the gifts.
| 2019-04-24T18:26:53 |
http://lawschool.mikeshecket.com/property/3-11-04.htm
|
0.999999 |
These mobile homes without sanitary facilities (and without running water) are equipped as follows: - Fully equipped kitchen with tableware, fridge, coffee machine, 4 burner gas hob, microwave, hood - 2 separate bedrooms with electric heaters and 2 single beds of 70x190 - Wooden terrace, garden furniture - Television - Free access to Wi-Fi.
- Equipped kitchen with tableware, fridge, coffee machine, 4 burner gas hob, hood, microwave - 1 Separate bedroom with 1 bed of 140x190 OR 2 beds of 80 x 190 - Electric heaters* - 1 bathroom with shower, washbasin and toilet - Garden furniture - Television - Free access to Wi-Fi.
- Lounge - Equipped kitchen with tableware, fridge, coffee machine, 4 burner gas hob, hood, microwave - 1 bedroom with 1 double bed of 140x190 - 1 bedroom with 2 single beds of 80 x 190 - Bathroom with shower, washbasin - Separate toilet - Electric heaters* - Garden furniture with parasol - Terrace or small garden - Television - Free access to Wi-Fi.
Mobil Home Premium 33m2 - - 1 seating area - 1 fully equipped kitchen area with crockery, fridge, coffee maker, kettle, toaster, gas hob 4 burners, hood, dishwasher and microwave - 1 bedroom with a double bed 160x190 and its private bathroom and separate toilet - 1 bedroom with 2 single beds 80 x 190 and its private bathroom and separate toilet Hotel formula: Beds made on arrival Towels provided - electric convectors in each room - 1 garden furniture - Terrace or private garden - TV included, flat screen and TNT - Free wifi access.
Premium COTTAGE 40 m2 - - 1 seating area - 1 fully equipped kitchen area with crockery, fridge, coffee maker, kettle, toaster, gas hob 4 burners, hood, dishwasher and microwave - 1 bedroom with a double bed 160x190 and its private bathroom and separate toilet - 1 bedroom with 2 single beds 80 x 190 - 1 bedroom with 2 single beds 80x 190 mezzanine Hotel formula: Beds made on arrival Towels provided - electric convectors in each room - 1 garden furniture - Terrace or private garden - TV included, flat screen and TNT - Free wifi access.
- 1 coin salon - 1 coin cuisine tout équipé avec vaisselle, frigo, cafetière électrique, table de cuisson au gaz 4 feux, hotte, micro-ondes - 1 chambre avec 1 lit double de 140x190 - 1 chambre avec 2 lits simples 80 x 190 - Salle d'eau avec douche, lavabo - W.C séparé - Convecteurs électriques* - 1 salon de jardin avec un parasol - Terrasse ou jardin privatif - Télévision incluse - Accès wifi gratuit.
No drain for waste water so had to lift a heavy container to empty it. Fresh water tap was not labelled.
Friendly reception and helpful staff. Toilets and showers kept clean.
Did not use the pool but it looked nice.
Piscine un peu petite dès qu'il y a du monde mais hors saison, ça le fait.
Belle terrasse agréable même au mois de novembre dès qu'il y a un rayon de soleil !
Proximité de certains espaces tente avec les bungalows; les personnes logeant dans un bungalow ne se rendant pas compte du bruit généré pour des campeur en tente juste à coté d'eux.
Grand espace dans lequel on peut mettre plusieurs tente et tout de même garer la voiture.
| 2019-04-23T14:02:55 |
https://book.campsited.com/france/pont-scorff/camping-ty-nenez
|
0.999263 |
I Love to make new *FRIENDS* DO U Know Friend that a simple "HELLO" canbe sweet one? The word "HELLO" means : H = "HOW are you? " E = "Every thing all right?" L = "Like to HEAR from you" L = "Love to SEE you soon" O = "Obviously, I MISS YOU "
working in london for 5 years, well established here. good income.
I am very honest, easy going, hard working, ambicious. I like reading , watching current afairs and sports.
I am a normal person looking for somone who loves me from deep inside.
Hello there, I am a very down to earth man seeking for friends all around the world. I am a very sporty type of person and I absolutely enjoy playing football at the most.
You're welcome to ask anything, so I'm not going to 'big' myself, i'd rather you decide.
I would like to know about how the rest of the world lives. I enjoy making friends around the world.
| 2019-04-24T06:41:50 |
https://www.indiancupid.com/en/men/uk/friendship/occupation/administration
|
0.999979 |
"The spirit of this country is totally adverse to a large military force," said Thomas Jefferson in an 1807 letter to Chandler Price. And what would the nation's third president and principal author of the Declaration of Independence say of today's American "global force" deployed across the earth?
America's "global military presence" is not hyperbole, but a literal fact, as a new Department of Defense study has found that US military personnel are stationed in literally every nation on the earth. The report bluntly confirms that, "The United States has military personnel in nearly every country in the world, ranging from two liaison officers in Fiji to tens of thousands from all of the service branches in Japan and Germany, according to the report."
But perhaps more important and equally absurd (when one considers the original beginnings of the humble "republic" envisioned by founders like Madison and Jefferson) is the fact that America's world-wide military presence is such that the Pentagon itself can't track how many US service members are deployed where. A new bombshell article in Stars and Stripes military newspaper is non-ironically headlined, "Report: 44,000 ‘unknown’ military personnel stationed around the world".
The Stripes article opens with a line that would seem straight out of the Onion, yet matter-of-factly states, "The U.S. military has more than 44,000 troops across the globe that the Pentagon claims it cannot track, according to a recent report."
The Pentagon's system for accounting troop deployment numbers, maintained by the Defense Manpower Data Center under the Office of the Secretary of Defense, lists up to 44,000 troops in a country location category labeled "Unknown".
The report has come to light after a series of recent Pentagon press briefings on military actions in the Middle East and Africa which left journalists scratching their heads. For example the latest DoD admission that there are "about 2,000 troops in Syria" - four times the previously acknowledged figure - came after a late October Pentagon presser wherein an Army spokesman told reporters that 4,000 troops were deployed to Syria, after which he awkwardly walked back the statement less than 30 seconds later, saying "I'm sorry I mispoke there - there are approximately 500 troops in Syria".
And when news broke of the October 4 ambush and deaths of four elite Green Beret soldiers in Niger, the immediate reaction voiced among congressional leaders and echoed generally in the media was: "we have troops in Niger?" - after which it came to light that the US has at least 6,000 military personnel spread across the African continent, though few knew of AFRICOM's (US Africa Command) dangerously burgeoning number of undisclosed daily missions.
“We are not at a point where we can give numbers other than those officially stated,” said Army Col. Rob Manning, quoted in Stripes. Though as Stripes acknowledges it appears that this is due more to the Pentagon simply not knowing where it has stationed thousands of troops rather than the officially stated reasons of "secrecy" and force protection from enemies. The Army spokesman further asserted, “Our commitment is to be as transparent as we can, within the constraints of operation security,” but it's also that case that, "The Pentagon acknowledged in a statement that it has no good way to track how many service members are stationed overseas, where they are and when they were there," according to Stripes.
The report also comes after President Trump unveiled a new Afghan strategy last summer, saying the US would not pull out of the now 16-year long war but would instead inject more troops into the country for increased anti-terror operations. According to recent statements from the Pentagon, the total number of US soldiers in Afghanistan has risen to about 14,000, though accounting procures have been especially unclear concerning the nation's longest running war.
When Secretary of Defense James Mattis was asked during a news conference last summer about troops levels in Afghanistan he said, “There’s a very strange accounting procedure I inherited... What I’m probably going to end up doing is out putting everyone into one thing and saying, ‘Here’s how many are really there now."
And none of this addresses the even greater numbers of civilian contractors in global hot spots where America is present. According to Stripes, "The United States maintains 23,659 contractors in Afghanistan and 4,609 in Iraq, according to a fourth-quarter 2017 report from the Defense Department."
Given that we learned this week of 44,000 troops stationed in locations the Pentagon itself deems "unknown", perhaps we are long overdue for a massive drawback of overseas forces and bloated military budgets? Of course if the Pentagon can't find them in the first place, it might be hard to bring them home.
| 2019-04-22T22:42:12 |
https://www.zerohedge.com/news/2017-12-09/dude-wheres-my-troops-pentagon-loses-track-44000-soldiers
|
0.999872 |
Q. If you use most of a previous paper (of your own) into a new one (new class), is this considered plagiarism?
Specifically, if I am in a different class, and the paper I have to write is of the same information as one I wrote in another semester, for a different class, is it okay to copy parts of it and use it in the new paper?
Plagiarizing yourself by copying and pasting large chunks of previously written material without citing it may be considered academically dishonest by most professors. The reason is, professors give you specific assignments with the hopes that you will do fresh work each time.
To ensure you are acting with academic honesty, you should ask your professor if you are allowed to quote or paraphrase your ideas from your old paper as long as you cite it. This will follow the same procedures as citing from any source, which means that the core of the new paper must differ from the old, and you can't use too large of portions of the old paper.
| 2019-04-23T23:05:36 |
http://ask.uta.edu/faq/67263
|
0.999999 |
For which task is the hypervisor responsible in a virtualized server environment?
A SAN administrator recently added a new zone to the VSAN 2 zoneset named zoneset2, but the server is still unable to access the shared storage. What configuration will resolve this issue?
Refer to the exhibit. NP uplink ports connected to the core NPIV switch do not come online and are stuck in an initializing state. What is the cause that NP Uplink ports on NPV edge switch are stuck in initializing state?
The upstream NP Ports and the downstream server F Ports on the NPV edge switch may not be in the same VSA.
The core switch has not been enabled for NPIV.
The upstream VSAN ports and the server ports are in full duplex mode.
Related VSAN identifier is not allowed on Virtual Port Channels toward the edge switch.
FCoE maps Fibre Channel onto Layer 2 Ethernet, Converging IP, and storage networks.
FCoE encapsulates native Fiber Channel traffic inside iSCSI packets.
FCoE encapsulates native Fiber Channel traffic inside GRE tunnels.
Data Center Bridging uses FCoE to transport IP traffic over native Fiber Channel.
FCoE encapsulates Native Fiber Channel inside IP packets.
Which statement describes what the customer is configuring in Exhibit 4?
Create a separate LACP for the vPC peer-keepalive link and add a Layer 3 interface to that VRF.
Create a separate VRF for the vPC peer-keepalive link and add a Layer 4 interface to that VRF.
Create a separate VRF for the vPC peer-keepalive link and add a Layer 3 interface to that portchannel.
Create a separate VRF for the vPC peer-keepalive link and add a Layer 3 interface to that VRF.
| 2019-04-22T03:00:28 |
http://www.70-742.com/free-2018aug-ensurepass-cisco-640-916-dumps-with-vce-and-pdf-11-20/
|
0.999997 |
Question: A Worker-communist Party of Iran (WPI) declaration states that the slogan of federalism proposed by some political parties is not in any way indicative of the existence of such a demand amongst people. In such a case, why was it necessary to issue this declaration?
Mansoor Hekmat: Not only does the slogan for federalism not have any place in people's minds and lives but ethnic identity and ethnocentrism also have no place amongst millions of people living in Iran. Unfortunately, however, this is not enough to immunise people against the tragedies and catastrophes that ethnocentrism and ethnocentrists can create. Look around the world and review the developments of the last two or three decades. What basis did clerical government and religious rule have in Iran in the twentieth century? Apparently, none. These were people who had dealt with this issue eighty years earlier. Mullahs were a disgraced parasitic segment and the subject of people's ridicule in cities and villages. Two generations before us, girls went to school unveiled. Music and cinema were an inherent part of people's lives. Despite this, however, religious insanity and ignorance and Islamic beasts rule in that society today. Yugoslavia was a modern, industrial, and by any contemporary standards, civilised society. Yugoslavs say that before these events, they did not remember their ethnicity and nationality. They still don't believe that this happened and don't know how it happened.
Since ethnicity and ethnocentrism don't have strong roots and foundations in Iranian society and the demand for federalism does not have any place amongst people, the main groups and movements in society don't move in that direction. The question is, however, how immune is society to ethnocentric provocation and how prepared is it beforehand to defend itself in the face of regressive and reactionary ethnocentrism's future assault? Nationalism and ethnocentrism are like viruses that exist in a hibernated state, become active under specific circumstances, and even sometimes cause epidemics. In contemporary Iranian society, nationalism and ethnocentrism are not a widespread political-cultural ailment, however, it has neither been uprooted nor is society immunised against it. The existence of these federalists and a backward bunch that have already, with utmost shamelessness and idiocy, started issuing ethnic identity cards for the people of Iran attests that this virus has not been uprooted yet and can, particularly in a changing and turbulent political situation, bring many calamities. This declaration has been issued to create this preparedness and immunity in society. Furthermore, it is a clear announcement to the nationalists and ethnocentrists and the half-witted intellectuals surrounding them that worker-communism will resolutely stand up to this regressiveness and disgrace these inhuman and anti-worker tendencies and groups. They cannot enslave and suppress people for twenty years with religion and then another twenty years with ethnicity and nationalism. We will not allow it.
Question: The WPI declaration directly relates the advent of the slogan for federalism to the Kurdish question. Could you explain this further?
Mansoor Hekmat: As I have said, labelling people along nationality and ethnicity, let alone along ethnic or national conflicts between various segments of society, is not a persistent, widespread, or all encompassing phenomenon in the country. Clearly, in the backward folklore of ethnic groups (any ethnic group), there is an ethnic egotism, chauvinism and xenophobia, etc. This is true in all countries, even in the most homogeneous 'nations'. Evidently, as long as capital and capitalism exists, the production of divisions amongst people will remain. Iran, too, is not an exception. But any observer who does not have an axe to grind will agree that Iranian society is not a society crippled with ethnic and nationalist conflict and intolerance for which some pundit need come up with the federalist antidote.
When you pay attention, you see that recently, this subject has been raised directly regarding the Kurdish question and the dealings and negotiations between the Kurdish Democratic Party of Iran (KDPI) and the nationalist opposition that are, according to the KDPI, due to take over the central government of tomorrow. The whole concern of the KDPI is to avoid being labelled secessionist, especially since it knows the brutal capacity of Iranian expansionist nationalists and the crimes they are willing to commit under the pretext of defending 'territorial integrity'. Consequently, the KDPI itself has come in the lead of defending 'territorial integrity'. In an interview, when Mr. Mullah Abdullah Hassanzadeh, the General Secretary of the KDPI, is asked about the Worker-communist Party of Iran position (recognition of Kurdistan's right to separate; holding a referendum to determine people's votes over secession or remaining within the framework of Iran as citizens with equal rights), he says: 'No, we are not secessionists'. In order to lay the next central government's fears at rest, he draws a very clear line with us. The KDPI wants autonomy and has concluded that it would be better to extend its demand to the entire country by presenting a formula that will not leave Kurdish nationalism alone vis-à-vis the nationalist central government. The federalism formula helps the KDPI to ask for autonomy without Kurdistan becoming an exception. Federalism means giving autonomy to 'all nations in Iran', including the Kurdish nation led by the KDPI. It doesn't concern them that the rest of the people of Iran, including a large segment in Kurdistan itself, do not necessarily define themselves with ethnic labels, such as Fars, Lor, Gilak, Afghan, Kurd, Arab, Baluch, and Turk.
On the other side of this reactionary wheeling and dealing stands the Iranian nationalist groups which see themselves faced with the Kurdish question as one of the main obstacles to establishing their future rule. They see in the slogan of federalism a window of opportunity to sidestep the Kurdish question, bypass the people's votes and wheel and deal with the KPDI.
The source of the debate on federalism is the opportunist calculations of several political parties regarding the Kurdish question. This in itself is not very worrying and not difficult to respond to with persistent communist and anti-nationalist publicity. Next to the main participants of this debate who are pursuing their material interests, however, we find, as usual, a group of people who because of their honourable journalistic or intellectual duties or their organisational needs, must say something according to the fad of the day and without any serious contemplation. Attesting to the bitter fate of the people of Yugoslavia and the Soviet Union, they are the most dangerous elements in such a situation. It is truly unbelievable that someone sees his child playing outside with the next-door neighbours' children from his window, then sighs, takes his pen, and foolishly divides the people of a 60 million populated country into 9 nationalities, 13 ethnic groups and 5 language groups and prescribes that they must divide cities and localities between them according to his chart and stand face to face against each other. If these foolish thoughts didn't have such bloody consequences, perhaps they would even be amusing for their listeners. They are, however, consciously or unconsciously paving the way for tomorrow's ethnic cleansing and mass executions. They are conspiring against the lives of those children playing outside the window. This is where the sensitivity of the issue lies.
This is not the world of 30 years ago. The information and electronic revolution and the appearance of worldwide media have turned lackey journalism and the engineering of opinion into a determining factor in the power struggle and political development of various countries. Even the most foolish and far fetched political tendencies and social prescriptions, if in the interests of a section of the ruling class in a broader class and international framework, will be propped up when needed and imposed on people's perceptions. The question goes far beyond the petty calculations of the KDPI and remainders of nationalist, republican or monarchist parties. They are pawns. We must make the entire game impossible.
I must add that as far as the Kurdish question is concerned, it is the very people of Kurdistan themselves who must decide about secession or remaining within the framework of Iran as citizens with equal rights in a free referendum. Only this decision, whatever it will be, is legitimate and principled. The Kurdish question is not about the share of the KDPI and other autonomy-seeking parties in the power structure and handing 'Kurdish rule' to them over people's heads. Any wheeling and dealings and negotiations by parties over Kurdistan have no legitimacy.
Question: Can it be said that this slogan is characteristics of the political parties and groups proposing it and must they be judged accordingly? Does this declaration change the WPI's relationship and position towards the parties and groups which advocate federalism?
Mansoor Hekmat: In my opinion, parties can be judged based on this slogan at the theoretical level. Just as being religious, national chauvinist, monarchist or opposing the equality of men and women is enough to call a group reactionary within the context of Iran's contemporary political history, in the same way, advocating the issuance of ethnic identity cards and the establishment of a government based on nationality and ethnicity is enough to call a group reactionary. If there is a question, it is whether the depths of the filth of the federalism slogan and its inhuman, anti-social and anti-worker character have been extensively recognised or not. Perhaps, not yet. Initially, for a period, this makes it necessary to strive to remove this formula from the programme and publicity of opposition parties. It is clear that from now on, any organisation that consciously and persistently defends federalism or the ethnic organisation of society must be considered reactionary.
Question: What tasks has this declaration put forth for WPI activists? What must be done vis-à-vis groups proposing federalism?
Mansoor Hekmat: Clearly, it is not possible nor is it necessary to list the activities here. The crux of the matter is that we will strongly and persistently make the real meaning of this slogan clear for people. The parties which advocate this slogan must feel the full force of our criticism wherever they go. We must disgrace ethnic and national politics and philosophy in the same way that religious philosophy and politics has been disgraced and become ostracized today. We must have, if not all of the working people, then at least a more active and dynamic section of the working class in Iran which is so sensitive to and consciously opposed to reactionary nationalism and ethnocentrism (the federalism slogan is only a fancy version of it) to the extent that no force can stop the advance of the worker and worker-communism by creating ethnic conflicts and atrocities and bringing about a scenario like what we witnessed in Yugoslavia. Our publicity and actions must make the emergence of indigenous Tudjmans, Karadzics, Mladics and Milosevics impossible. In a word, we must heighten people's attention and make all the ethnocentrists and their sponsors of tomorrow understand that they face an uncompromising worker-communism, which has no illusions.
This was first published in June 1996 in International 21 in Persian. The English version is a reprint from WPI Briefing.
| 2019-04-24T13:48:58 |
http://hekmat.public-archive.net/en/0930en.html
|
0.999522 |
Donizetti's Anna Bolena takes place in England during the 16th century.
Within the queen's chambers at the castle, Queen Anna has become aware her star is dimming as King Henry VIII is falling in love with another woman. Queen Anna calls in her confidante and lady-in-waiting, Jane Seymour, to vent her frustrations. Queen Anna's countenance and sad demeanor have affected all those in her chamber, so she orders her page, Smeton, to play a song on the harp to cheer everyone's mood. After the song, everyone leaves, except for Jane. Not long after, King Henry VIII enters and tells Jane that his love for her grows stronger and that she will be at his side at the altar in due time. When King Henry leaves, Jane grows weary of her decision and how it will affect the queen. However, she decides that their affair has gone too far to stop it now.
The next day, Queen Anna's brother, Lord Rochefort, rides through Richmond Park and happens across Queen Anna's former lover, Lord Richard Percy. Surprised, Rochefort asks Percy why he has returned. Percy replies that it was the king himself who called him back from exile. Percy questions Rochefort about Queen Anna's happiness after he heard rumors of the deteriorating state of their relationship. Rochefort somewhat dodges the question but tells him that love is often never a part of royal marriages.
Back in Queen Anna's chambers, Smeton, who has fallen in love with the queen, has stolen a small portrait of her and has come back to return it. Before he can put back the portrait, he hears a noise outside of her chamber and hides behind a screen. Queen Anna enters with her brother, Rochefort. Rochefort asks Queen Anna to give a moment of her time for Percy. She agrees, which catches Smeton's attention. He eavesdrops on their conversation since he cannot escape without being caught. When Rochefort leaves, Percy enters the room. Percy tells Queen Anna that he knows she is unhappy. She tells him that the king has grown to loathe her. Percy confesses he still has strange feelings for her and asks her to leave with him. When she refuses, Percy draws his sword and attempts to kill himself. When Queen Anna shouts, Smeton thinks that Percy is attacking her, so he jumps out from behind the screen. Percy turns his sword to Smeton and the two start to fight. Not long into their fight, King Henry VIII and his men burst into the room.
The king orders their arrest, but Smeton pleads Queen Anna's innocence. He tells the king that he can stab him in the heart if he is lying and rips open his tunic. When he does, the small portrait of Anna falls to the king's feet. Happily, the king finds proof for condemning Queen Anna and sends them all to jail.
King Henry VIII has imprisoned Queen Anna in her London apartment. Jane arrives to help the queen escape her impending execution, and tells her that if she confesses her love for Percy, the king will grant her freedom, giving him cause for divorce. This is all the king wants. Queen Anna, loyal to her duties and marriage oath, refuses and wishes her successor a crown of thorns. Jane, full of guilt, reveals that she is the king's secret lover. Queen Anna is furious but finally calms down when Jane convinces her that the king is the guilty one.
Hoping to save the queen, Smeton falsely testifies that he has had relations with her. Unknowingly, Smeton sets the queen's fate in stone. Percy and Anna are brought into the antechamber separately. Percy claims that he and Anna were married prior to her marriage with the king, but the king does not believe him. Anna pleads with the king to keep this news from the public eye but is ready to give up her life. Jane fights for Queen Anna's life, but the king ignores her. King Henry VIII dismisses everyone and the council decides to execute all three of them as well as dissolving the king's marriage to Anna.
Within Queen Anna's cell, her stress and grief have caused her to go mad. She deliriously recalls her past memories with Percy. Percy and Smeton are brought into her cell and Smeton begs for her forgiveness. She looks down at Smeton in bewilderment and asks him why he isn't playing his music. Moments later, the sound of the cannons is heard, signaling the king's new marriage. Queen Anna snaps out of her temporary madness and curses the king. Full of pride, she marches to her execution with her head held high.
| 2019-04-21T20:35:59 |
https://www.thoughtco.com/anna-bolena-synopsis-723734
|
0.998904 |
The global demography project is an effort to generate consistent, spatially referenced global population data sets for global and regional environmental analysis, demographic research, etc. This project has been supported by the Consortium for International Earth Science Information Network (CIESIN) , the Environmental Systems Research Institute (ESRI) and NCGIA. Output from the first stage of the project includes a set of raster data sets of population density and total population figures. For each, a smoothed and an unsmoothed version exists. The smoothed images were produced using Waldo Tobler's pycnophylactic interpolation. These images have a resolution of 5 minutes and are based on source population and boundary data for more than 15,000 administrative units. The project is described in detail in NCGIA Technical Report 95-6 "The Global Demography Project" by Waldo Tobler, Uwe Deichmann, Jon Gottsegen, and Kelly Maloy. The full report including maps and illustrations can be obtained from NCGIA Publications [email protected]. The raster images are distributed by CIESIN and can be viewed on CIESIN's website.
More information about the global demography project.
Several population databases for Africa have been compiled. A first, medium resolution database for Africa had been compiled for a related project on African farming systems which was supported by the Rockefeller Foundation and NCGIA.
A more recent, updated database consists of boundaries for about 4500 administrative units and population estimates for 1960, 70, 80 and 90. Again regular raster grids of population density have been developed from these data sets which are distributed from the UNEP/GRID Sioux Falls website. This work was supported by UNEP/GRID Sioux Falls and the World Resources Institute.
to contribute to further development of global databases relevant to agricultural research.
The third objective has led to a number of initiatives to improve global demographic data holdings. This includes updating of data holdings for Latin America by CIAT , and an update and enhancement of the Asian data compiled for the global demography project by NCGIA and UNEP/GRID-Geneva. The documentation and data sets for are available at UNEP/GRID Geneva, Switzerland and UNEP/GRID Sioux Falls, South Dakota, USA.
Note: the surfaces are slightly smoothed for visual presentation; the actual resolution is higher.
Also, a review paper on population databases and modeling has been published as NCGIA Technical Report TR 96-3.
For more information about the UNEP/CGIAR Initiative contact GRID Arendal ([email protected]) .
Please note, the URL is only found on archive.org, some of the links on this page may not work.
| 2019-04-20T14:56:07 |
http://www.ncgia.ucsb.edu/pubs/gdp/pop.html
|
0.995299 |
Systems and methods for measuring macromolecular proton fraction in a subject are provided. A nuclear magnetic resonance apparatus applies a magnetic field to a body region on the subject, and radiofrequency modes are applied to the body region as well. Each radiofrequency mode delivers a plurality of radiofrequency pulses separated by time delays, wherein at least one of the radiofrequency modes causes suppression of signal components from an unwanted tissue, and at least one of the radiofrequency modes causes magnetization exchange between water and macromolecules in tissues in the body region. Amplitudes corresponding to magnetic signals received from the body region are measured and macromolecular proton fraction based on the amplitudes can be calculated.
This invention was made with U.S. government support under Grant No. R21EB016135, awarded by the National Institutes of Health. The U.S. government has certain rights in the invention.
1. A method for measuring macromolecular proton fraction in a subject comprising: applying, via a nuclear magnetic resonance apparatus, a magnetic field to a body region on the subject; applying a plurality of radiofrequency modes to the body region, wherein each radiofrequency mode comprises delivering a plurality of radiofrequency pulses separated by at least one time delay such that (i) at least one of the plurality of radiofrequency modes causes suppression of signal components from an unwanted tissue in the body region, and (ii) at least one of the plurality of radiofrequency modes causes magnetization exchange between water and macromolecules in tissues in the body region; measuring amplitudes corresponding to a plurality of magnetic signals received from the body region; and calculating macromolecular proton fraction in the body region based on the measured amplitudes.
2. The method of claim 1, wherein the body region encompasses one or both of the brain and the spinal cord of the subject.
3. The method of claim 2, wherein an unwanted tissue is a tissue not belonging to the brain.
4. The method of claim 3, wherein the tissue not belonging to the brain is one or more of: (i) a fat tissue; (ii) a muscle tissue; and (iii) cerebrospinal fluid.
5. The method of claim 2, wherein delivering the plurality of radiofrequency pulses separated by at least one time delay causes one or more of the following: (i) a saturation of a nuclear magnetization of fat protons outside the brain; (ii) a saturation of a nuclear magnetization of muscle protons outside the brain; (iii) a saturation of a nuclear magnetization of cerebrospinal fluid protons inside the brain; (iv) a variable excitation of a nuclear magnetization of protons of the brain; and (v) a variable saturation of a nuclear magnetization of macromolecular protons in the brain parenchyma.
6. The method of claim 2, wherein measuring amplitudes corresponding to the plurality of magnetic signals is performed after varying one or more of the following: (i) the excitation of a nuclear magnetization of protons of the brain; (ii) the saturation of the nuclear magnetization of macromolecular protons in the brain parenchyma; and (iii) a duration of the time delay separating the radiofrequency pulses.
7. The method of claim 1, wherein the subject is a human subject and is one of: (i) an adult human subject; (ii) an adolescent human subject; (iii) an infant human subject; and (iv) a newborn human subject.
8. The method of claim 1, further comprising: comparing the macromolecular proton fraction values for the body region with a control.
9. The method of claim 1, wherein measuring amplitudes corresponding to a plurality of magnetic signals further comprises varying one or more of the following: (i) an excitation of a nuclear magnetization of protons in the body region; (ii) a saturation of a nuclear magnetization of macromolecular protons in the body region; and (iii) a duration of the at least one time delay separating the plurality of radiofrequency pulses.
10. The method of claim 1, further comprising: determining a myelin content in the body region based on the macromolecular proton fraction.
11. A magnetic resonance apparatus comprising: a magnet configured to apply a magnetic field to a body region on a subject; a radiofrequency pulse generator configured to deliver radiofrequency pulses to the body region; and a controller comprising a processor and a non-transitory computer-readable medium configured to store program instructions thereon executable by the processor to cause the controller to perform functions comprising: applying the magnetic field to the body region; applying a plurality of radiofrequency modes to the body region, wherein each radiofrequency mode comprises delivering the plurality of radiofrequency pulses separated by at least one time delay, such that the said plurality of radiofrequency pulses causes suppression of signal components from an unwanted tissue in the body region for at least one of the plurality of radiofrequency modes and magnetization exchange between water and macromolecules in tissues in the body region for at least one of the plurality of radiofrequency modes; measuring amplitudes corresponding to a plurality of magnetic signals received from the body region; and calculating macromolecular proton fraction values of the body region based on the measured amplitudes.
12. The magnetic resonance apparatus of claim 11, wherein an unwanted tissue of the subject is one or more of: (i) a fat tissue; (ii) a muscle tissue; and (iii) cerebrospinal fluid.
13. The magnetic resonance apparatus of claim 11, wherein the body region encompasses one or both of the head and the spinal cord of the subject.
14. A method for measuring macromolecular proton fraction in a subject, comprising: applying a magnetic field to a body region of a subject; applying a first plurality of radiofrequency pulses at a nuclear magnetic resonance frequency so that each pulse of the first plurality of radiofrequency pulses causes nutation of a nuclear magnetization in the body region by a first flip angle; recording a first magnetic resonance signal received; applying a second plurality of radiofrequency pulses at the nuclear magnetic resonance frequency so that each pulse of the second plurality of radiofrequency pulses causes nutation of a nuclear magnetization in the body region by a second flip angle different from the first flip angle; recording a second magnetic resonance signal received; applying a plurality of saturation radiofrequency pulses at frequency different from the nuclear magnetic resonance frequency so that each pulse of the plurality of saturation radiofrequency pulses causes saturation of a nuclear magnetization of macromolecular protons in the body region; applying a third plurality of radiofrequency pulses at the nuclear magnetic resonance frequency so that each pulse of the third plurality of radiofrequency pulses causes nutation of a nuclear magnetization in the body region by a third flip angle; recording a third magnetic resonance signal received; and calculating the macromolecular proton fraction by solving an equation relating the first magnetic resonance signal, the second magnetic resonance signal, and the third magnetic resonance signal to the macromolecular proton fraction.
15. The method of claim 14, further comprising: spatially encoding the first, the second, and the third magnetic resonance signals; and constructing a map comprising a spatial distribution of the macromolecular proton fraction values.
16. The method of claim 15, wherein the body region is a brain of the subject.
17. The method of claim 16, further comprising: suppressing signal components from tissues not belonging to the brain by magnetization preparation.
18. The method of claim 17, wherein magnetization preparation includes one of the following: (i) inversion-recovery; (ii) partial inversion-recovery; and (iii) spectrally-selective saturation.
19. A method of mapping macromolecular proton fraction using only three source images, comprising: reconstructing longitudinal relaxation time and proton density maps from two variable flip angle source images; synthesizing a synthetic reference image from the longitudinal relaxation time and proton density maps; and reconstructing a macromolecular proton fraction map from a magnetization transfer (MT) weighted image normalized to the synthetic reference image and the longitudinal relaxation time map.
20. The method of claim 17, wherein reconstructing the longitudinal relaxation time and proton density maps from two variable flip angle source images further comprises processing the two variable flip angle source images with corrections for magnetic and radiofrequency field inhomogeneities, and wherein synthesizing a synthetic reference image further comprises synthesizing the synthetic reference image from the longitudinal relaxation time map, the proton density map, and a radiofrequency field inhomogeneity map.
21. The method of claim 17, wherein the method is used to determine the myelin content in a tissue of a human subject.
22. The method of claim 17, wherein the method is used to monitor gray matter myelination.
This application claims priority to U.S. Provisional Patent Application Ser. No. 62/074,480 filed on Nov. 3, 2014, which is hereby incorporated by reference in its entirety.
Myelination, which is the production of the myelin sheath, can be viewed as a morphological criterion of the functional maturity of neural conduction in the brain. Cerebral myelination begins before birth, and is not completed until late in adolescence.
The assessment of cerebral myelination is useful for a number of functions. Cerebral myelination at various life stages may be observed and monitored. Additionally, a number of neurological diseases, where the loss of myelin slows or blocks nerve conduction, may be diagnosed, monitored, and be subject to prognostic assessment through observing and measuring brain and spinal cord myelination. Such diseases include multiple sclerosis (MS), brain trauma, stroke, and many other conditions.
Magnetic resonance imaging (MRI) is a widely used medical imaging modality that provides soft-tissue differentiation to assess cerebral myelination in a subject in vivo. However, clinical MRI scanners, especially the superconducting magnet and magnetic field gradient hardware of MRI scanners, are too costly to use in population studies or in low-resource healthcare facilities.
In accordance with the present invention, a system and a method are defined for measuring macromolecular proton fraction in a subject. In one embodiment, the method may comprise applying, via a nuclear magnetic resonance apparatus, a magnetic field to a body region on the subject and applying a plurality of radiofrequency (RF) modes to the body region. Each RF mode comprises delivering a plurality of RF pulses separated by at least one time delay such that (i) at least one of the plurality of RF modes causes suppression of signal components from an unwanted tissue in the body region, and (ii) at least one of the plurality of RF modes causes magnetization exchange between water and macromolecules in tissues in the body region. Amplitudes are then measured corresponding to a plurality of magnetic signals received from the body region, and macromolecular proton fraction in the body region is calculated based on the measured amplitudes. In some embodiments, the body region may encompass the brain or the spinal cord in the subject. An unwanted tissue may be a tissue that does not belong to the brain. The unwanted tissue may be a fat tissue, a muscle tissue, or a cerebrospinal fluid. The subject may be an adult, an adolescent, an infant, or a newborn infant.
Delivering the plurality of RF pulses separated by at least one time delay may cause one or more of the following: (i) a saturation of a nuclear magnetization of fat protons outside the brain; (ii) a saturation of a nuclear magnetization of muscle protons outside the brain; (iii) a saturation of a nuclear magnetization of cerebrospinal fluid protons inside the brain; (iv) a variable excitation of a nuclear magnetization of protons of the brain; and (v) a variable saturation of a nuclear magnetization of macromolecular protons in the brain parenchyma.
In some embodiments, measuring amplitudes corresponding to the plurality of magnetic signals is performed after varying one or more of the following: (i) the excitation of a nuclear magnetization of protons of the brain; (ii) the saturation of the nuclear magnetization of macromolecular protons in the brain parenchyma; and (iii) a duration of the time delay separating the RF pulses.
In some embodiments, measuring amplitudes corresponding to a plurality of magnetic signals further comprises varying one or more of the following: (i) an excitation of a nuclear magnetization of protons in the body region; (ii) a saturation of a nuclear magnetization of macromolecular protons in the body region; and (iii) a duration of the at least one time delay separating the plurality of RF pulses.
Myelin content in the body region may be determined based on the macromolecular proton fraction.
In another embodiment, a magnetic resonance apparatus is provided and comprises a magnet configured to apply a magnetic field to a body region on a subject, an RF pulse generator configured to deliver RF pulses to the body region, and a controller comprising a processor and a non-transitory computer-readable medium configured to store program instructions thereon executable by the processor to cause the controller to perform functions comprising: applying the magnetic field to the body region, applying a plurality of RF modes to the body region, wherein each RF mode comprises delivering the plurality of RF pulses separated by at least one time delay, such that the said plurality of RF pulses causes suppression of signal components from an unwanted tissue in the body region for at least one of the plurality of RF modes and magnetization exchange between water and macromolecules in tissues in the body region for at least one of the plurality of RF modes, measuring amplitudes corresponding to a plurality of magnetic signals received from the body region, and calculating macromolecular proton fraction values of the body region based on the measured amplitudes.
These as well as other aspects and advantages of the synergy achieved by combining the various aspects of this technology, that while not previously disclosed, will become apparent to those of ordinary skill in the art by reading the following detailed description, with reference where appropriate to the accompanying drawings.
FIG. 17 depicts a graph of Pearson's correlation between whole-head MPF measurements and a subject's age after logarithmic transformation of data in accordance with at least one exemplary embodiment.
In the following detailed description, reference is made to the accompanying figures, which form a part thereof. In the figures, similar symbols typically identify similar components, unless context dictates otherwise. The illustrative embodiments described in the detailed description, figures, and claims are not meant to be limiting. Other embodiments may be utilized, and other changes may be made, without departing from the spirit or scope of the subject matter presented herein. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are explicitly contemplated herein.
Macromolecular proton fraction (MPF) is a biophysical parameter determining the efficiency of magnetization transfer (MT) between water and macromolecules in tissues. MPF is a sensitive and accurate non-invasive measure of myelin in neural tissues and is based on the strong linear correlation between MPF and histologically determined myelin density. Thus, MPF may be considered to be used as a myelin biomarker in clinical applications. The present disclosure provides a time-efficient, high-resolution method for in vivo MPF measurements in the human brain and spinal cord.
While cerebral myelination can be assessed using modern MRI methods, clinical MRI scanners are too costly to use in population studies or in locations with low financial resources, such as rural healthcare facilities and developing countries. The present disclosure is based on a whole-brain measurement of the specific biophysical parameter, MPF, from a non-localized magnetic resonance signal. By obtaining a whole-brain measurement from a non-localized magnetic resonance signal, the more expensive components present in clinical MRI scanners can be eliminated. Thus, the methods and systems described herein allow for the establishment of a trajectory of brain development across a wide age range and can be reliably implemented in low-resource settings.
FIG. 1 depicts a schematic of an exemplary system 100 in accordance with at least one embodiment. The system 100 may be used, among other things, to measure MPF in a subject. Thus, the system 100 may be used on a subject in vivo. As referenced herein, a subject may be a human subject, and may be an adult human subject, an adolescent human subject, an infant human subject, or a newborn human subject.
In FIG. 1, an MRI system is shown as system 100. The system 100 may include a magnet 110, an RF pulse generator 115, a controller 120, and a display 125. A sample to be imaged may be placed through the access 111 to be properly positioned for exposure to the magnet 110.
The system 100 provides for a non-localized, non-invasive measurement of the myelin content of a body region. The system 100 may implement low-cost hardware, avoiding the more expensive components that are generally present in clinical MRI equipment (e.g., a superconducting magnet and magnetic field gradient hardware). Advantageously, even with low-cost hardware, the system 100 can be used to quickly obtain a high-resolution macromolecular proton fraction of a sample. With the exception of the magnet 110, the electronic components of the system 100 enable a compact tabletop design. The system 100 includes low power consumption and may have a total weight less than 100 kg, enabling installation in any low-resource setting or manufacturing the system as a mobile unit on a vehicle platform.
In FIG. 1, the magnet is shown within a standard MRI structure providing the access 111 through which a sample can be placed for exposure to a magnetic field and subsequent analysis.
In one example embodiment, the magnet 110 is configured to produce a static and substantially uniform magnetic field encompassing the sample, which may include a body region of a subject. While in some embodiments the magnet 110 may be a superconducting magnet, the system 100 does not need implementation of a magnetic field gradient for spatial localization of a signal; thus, a low-cost, low-field magnet can achieve the desired results as the magnet 110. For example, the magnet 110 may be a low-field Halbach array magnet having a 0.1-0.2 Tesla magnetic field induction and producing a homogeneous magnetic field volume of, for example, a standard adult head size. In certain embodiments, the magnet 110 may also contain an appropriate transmit-receive RF coil and a dedicated pulse programming module, facilitating implementation of the method 200 of FIG. 2 below.
The RF pulse generator 115 is configured to deliver RF pulses to a body region of a subject.
The controller 120 may be a programmable radiofrequency controller (PRC) comprising a processor coupled to a receiver, data storage, and logic. These elements may be coupled by a system or bus or other mechanism. The processor may include one or more general-purpose processors and/or dedicated processors, and may be configured to perform an analysis on received and/or stored data. The controller 120 may be further configured to measure amplitudes corresponding to a plurality of magnetic signals received from the body region and to calculate macromolecular proton fraction in the body region based on the measured amplitudes.
An output interface of the controller 120 may be configured to transmit output from the controller 120 to the display 125. The display 125 may be any interface configured to display an output, such as data, maps, graphs, and images, for example.
In one embodiment, the system 100 may measure an MPF in the brain of a subject and compare the MPF with a control. In one example embodiment, a control includes a range of values of the MPF that is normal for a certain age of the subject. Such a comparison may be used to diagnose a demyelinating disease, for example.
In another embodiment, the system 100 may measure an MPF in a body region of a subject and determine a myelin content for the body region.
FIG. 2 depicts a simplified flow diagram of an example method 200 that may be carried out to measure MPF in a subject, in accordance with at least one embodiment. Method 200 presents an embodiment of a method that, for example, could be used with the system 100.
In addition, for the method 200 and other processes and methods disclosed herein, the flowchart shows functionality and operation of one possible implementation of the present embodiments. In this regard, each block may represent a module, a segment, or a portion of program code, which includes one or more instructions executable by a processor for implementing specific logical functions or steps in the process. The program code may be stored on any type of computer readable medium, for example, such as a storage device including a disk or hard drive. The computer readable medium may include a physical and/or non-transitory computer readable medium, for example, such as computer-readable media that stores data for short periods of time like register memory, processor cache and Random Access Memory (RAM). The computer readable medium may also include non-transitory media, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. The computer readable medium may be considered a computer readable storage medium, a tangible storage device, or other article of manufacture, for example. Alternatively, program code, instructions, and/or data structures may be transmitted via a communications network via a propagated signal on a propagation medium (e.g., electromagnetic wave(s), sound wave(s), etc.).
The method 200 allows for determining MPF in a body region of a subject. The method 200 may be used to quantitatively characterize brain myelination and to observe brain development across a wide age range, from early infancy to adulthood. The method 200 may additionally be used in infancy or childhood to predict capacities thought to contribute to adult productivity and intelligence.
Initially, the method 200 includes applying, via a nuclear magnetic resonance apparatus, a magnetic field to a body region on a subject, at block 210. The magnetic field may be delivered from a nuclear magnetic resonance apparatus comprising a magnet, such as the MRI device with the magnet 110 of FIG. 1, in one example embodiment. In operation, a subject is positioned at a designated location within the MRI device (e.g., through the access 111 of FIG. 1) to allow for a body region to be exposed to the magnet 110. The magnet 110 then emits a magnetic field, which is applied to the body region of the subject. The body region may comprise the brain and/or the spinal cord in some embodiments, and is observed in vivo.
The method 200 then includes applying a plurality of RF modes to the body region, at block 220. Each RF mode comprises delivering a plurality of RF pulses separated by at least one time delay such that (i) at least one of the plurality of RF modes causes suppression of signal components from an unwanted tissue in the body region, and (ii) at least one of the plurality of RF modes causes magnetization exchange between water and macromolecules in tissues in the body region.
The plurality of RF pulses may comprise a sequence, and may be pre-programmed into a computing system, such as the controller 120 of FIG. 1. The plurality of RF pulses suppresses signals from unwanted tissues, for example non-brain tissues, without the use of magnetic field gradients, while allowing detection of the MT effect on the brain water signal. Specifically, the dual-inversion preparation module with optimized inversion times is applied to suppress signal contributions from non-brain tissues such as cerebrospinal fluid (CSF) and subcutaneous fat. Contributions from the muscle tissue and scalp skin can also be suppressed using spin-echo signal readout with a sufficiently long echo time, because these tissues have much shorter relaxation times (T.sub.2) than brain white and gray matter. To sensitize this sequence to the MT effect, a series of off-resonance saturation pulses can be applied through sequence delays. A reference measurement for signal normalization may be performed using the same sequence without saturation pulses. A saturation-recovery sequence with the same preparative modules may be used to measure global brain longitudinal relaxation time (T.sub.1).
The method 200 then includes measuring amplitudes corresponding to a plurality of magnetic signals received from the body region, at block 230.
The method 200 includes calculating MPF in the body region based on the measured amplitudes, at block 240. Whole-brain MPF is calculated using the model of the pulsed MT effect, with modifications taking into account the effect of prepared RF pulses. Each signal measurement takes about 10-20 seconds (with phase cycling) due to a high signal-to-noise ratio that can be achieved for the whole-brain water signal without magnetic field gradients, resulting in the total signal measurement time comprising about three minutes.
In one exemplary embodiment, the method 200 operates by applying three RF modes generated in any order by a programmable RF controller of a magnetic resonance apparatus. A schematic diagram 300 of a plurality of RF amplitudes corresponding to a plurality of RF pulses generated in each mode is depicted in FIG. 3. In the first mode, identical on-resonance RF pulses with a flip angle .alpha..sub.1 and duration t.sub.on separated by a delay t.sub.off are applied to the body region until the magnetization reaches the steady state, and a steady-state magnetic resonance signal S.sub.1 is recorded. In the second mode, a similar sequence of RF pulses with a flip angle .alpha..sub.2 is applied, and a signal S.sub.2 is recorded. In the third mode, two RF pulses are applied in rapid succession: a first pulse is an off-resonance magnetization transfer saturation pulse with an offset frequency .DELTA..sub.mt, flip angle .alpha..sub.mt, and duration t.sub.mt; and a second pulse is an on-resonance pulse with a flip angle .alpha..sub.3 and duration t.sub.on. The pulse blocks are separated by a delay t.sub.off that is adjusted to limit the specific absorption rate in an object and may or may not be equal to the delay t.sub.off applied in the first and the second modes. Similar to the first and the second mode, the sequence in the third mode is repeated until the magnetization reaches the steady state, and a signal S.sub.3 is recorded after an on-resonance pulse.
where b.sub.1(t) is the pulse envelope function normalized to the area of the rectangular pulse with the same duration and peak amplitude. As described in Yarnykh V. L., Fast Macromolecular Proton Fraction Mapping From a Single Off-Resonance Magnetization Transfer Measurement, Magnetic Resonance Medicine 2012; 68(1): 166-178, equation can be iteratively solved relative to MPF with appropriate constraints applied to other model parameters. In the described exemplary embodiment, the signal S.sub.3 is normalized to the parameter PD and is used as an input variable together with T.sub.1 to compute MPF in a body region. In the described exemplary embodiment, a signal from a fatty tissue can be effectively removed by either applying water-selective excitation pulses or repeating each radiofrequency mode twice with a delay t.sub.off chosen such that water and fat signals appear either in the same or in opposite phase. In the last modification, complex summation of two signals results in cancellation of an unwanted signal from fat.
In another exemplary embodiment, the method 200 operates by applying five radiofrequency modes generated in any order by a programmable radiofrequency controller of a magnetic resonance apparatus. A schematic diagram 400 of radiofrequency amplitudes corresponding to a plurality of radiofrequency pulses generated in each mode is depicted in FIG. 4. In the first mode, a fully relaxed spin-echo signal S.sub.1 is recorded from an object by applying a radiofrequency pulse with 90.degree. flip angle followed by a radiofrequency pulse with 180.degree. flip angle. In the second mode, a 180.degree. inversion pulse is executed prior to the spin-echo module with an inversion delay TI.sub.csf chosen to null a signal from cerebrospinal fluid, and a spin-echo signal S.sub.2 is recorded. In the third mode, a 180.degree. inversion pulse is executed prior to the spin-echo module with an inversion delay TI.sub.fat chosen to null a signal from fatty tissue, and a spin-echo signal S.sub.3 is recorded. In the fourth mode, two 180.degree. inversion pulses are executed prior to the spin-echo module with two inversion delays TI.sub.1 and TI.sub.2 chosen to simultaneously null signals from cerebrospinal fluid and fatty tissue, and a spin-echo signal S.sub.4 is recorded. In the fifth mode, a continuous train of off-resonance magnetization transfer saturation pulses with an offset frequency .DELTA..sub.mt, flip angle .alpha..sub.mt, duration t.sub.mt, and inter-pulse delay t.sub.off is applied prior to the spin-echo module, and signal S.sub.5 is recorded.
Quantities Q.sub.1, Q.sub.2, and Q.sub.3 are independent of signal contributions from fat and cerebrospinal fluid and represent only signal parameters associated with a brain tissue. Equations and can be iteratively solved by any non-linear fitting algorithm relative to parameters PD.sub.b and T.sub.1b. Subsequently, PD.sub.b and T.sub.1b values are supplied into equation , which is iteratively solved relative to MPF similar to equation . In summary, this exemplary embodiment demonstrates that MPF in a brain tissue of an object can be determined from a non-localized magnetic resonance signal and contributions to MPF from non-brain tissue can be effectively eliminated.
Thus, image-based MPF mapping is converted into a non-imaging method, where the measured signals are collected from the entire brain parenchyma, while unwanted signal components from non-brain tissues are suppressed by the appropriate RF pulse sequence.
The systems and methods described above may be used to study myelination in a body region of a subject, and may be used to study a subject at risk of any brain tissue pathology, including but not limited to multiple sclerosis, brain trauma, stroke, brain degeneration, as well as neurological disorders such as Alzheimer disease and Huntington disease. Additionally, the systems and methods described above may be used to monitor myelination in infants and predict capacities thought to contribute to adult productivity and intelligence.
In a recent study, a high-resolution whole-brain MPF mapping method that utilizes a minimal number of source images to reduce scan time was conducted. The study is discussed in the conference proceedings, Yarnykh V. L., Fast High-Resolution Whole-Brain Macromolecular Proton Fraction Mapping Using a Minimal Number of Source Images, Proceedings of the 22th Annual Meeting of ISMRM-ESMRMB, Milan, Italy, 2014; p. 3335, which is incorporated herein by reference in its entirety.
The method applied in the study was a single-point method, wherein an MPF map was obtained by iteratively solving the pulsed magnetization transfer matrix equation using the Gauss-Newton method, with standardized constraints for non-adjustable two-pool model parameters and corrections for B.sub.0 and B.sub.1 inhomogeneities. A single-point MPF measurement was obtained from a spoiled gradient echo (GRE) MT-weighted image, a reference GRE image for data normalization, and an independently acquired T.sub.1 or R.sub.1 (wherein R.sub.1=1/T.sub.1) map.
A time efficient approach to measure T.sub.1 or R.sub.1 is the two-point variable flip-angle (VFA) method, employing a T.sub.1-weighted and a proton density (PD)-weighted GRE image.
The proposed algorithm synthesizes a reference GRE image from T.sub.1 or R.sub.1 and PD maps with the repetition time and VFA corresponding to those for the MT-weighted image according to the Ernst equation. If VFA data are processed with B.sub.1 correction, a B.sub.1 map can also be used to calculate a synthetic reference GRE image with actual flip angles in each voxel. Thus, if a synthetic reference GRE image is used, only three source images (MT, T.sub.1, and PD-weighted) are needed for MPF mapping.
In the study, MPF maps were obtained based on data from 6 healthy subjects (1 female, 5 males, age range 32-63 years). Images were acquired on a 3T (Philips.RTM. Achieva) scanner with an 8-channel head coil. PD and T.sub.1-weighted GRE images were acquired with repetition time (TR)=21 ms and FA=4.degree. and 25.degree., respectively. MT-weighted GRE images were acquired with TR=28 ms, FA=10.degree., and an off-resonance saturation pulse with single-lobe sinc-gauss shape, an offset frequency of 4 kHz, an effective saturation FA 560.degree., and a duration of 12 ms. Reference GRE images (TR=28 ms, FA=10.degree.) without saturation were also obtained for comparison purposes. All images were acquired in the 3D mode with non-selective excitation, FOV=240.times.240.times.180 mm.sup.3, and an isotropic voxel size of 1.25.times.1.25.times.1.25 mm.sup.3. To gain signal-to-noise ratio (SNR), a dual-echo acquisition (first/second echo time TE.sub.1/TE.sub.2=2.3 ms/6.9 ms) was used with summation of individual echo images during reconstruction. Dual-echo B.sub.0 maps and actual flip-angle imaging (AFI) B.sub.1 maps were acquired with voxel sizes of 2.5.times.2.5.times.2.5 mm.sup.3 and 2.5.times.2.75.times.5.0 mm.sup.3, respectively. Parallel imaging was used in both phase encoding directions with acceleration factors of 1.5 (anterior-posterior) and 1.2 (left-right). Scan times were 4 min 21 s for the T.sub.1 and PD-weighted images, 5 min 48 s for the MT-weighted image and the reference image, 2 min 8 s for the B.sub.0 map, and 3 min 26 s for the B.sub.1 map, thus resulting in an about 20 min acquisition time for the entire protocol without the reference image.
FIG. 5 depicts an image 500 comprising reformatted views of a high-resolution 3D MPF map, in accordance with an example embodiment. The image 500 was obtained with the synthetic reference image method described above. The image 500 demonstrates sharp white matter to gray matter (WM/GM) contrast and clear visualization of anatomical details of the brain.
FIG. 6 depicts a graph 600 illustrating voxel fraction percentage plotted over MPF percentage, in accordance with at least one embodiment. In the graph 600, group whole-brain MPF histograms were computed from MPF maps reconstructed using acquired (MPF.sub.acq) and synthetic (MPF.sub.synt) reference images and were averaged across the 6 healthy subjects. This group whole-brain MPF histogram demonstrates a very close agreement between the acquired and the synthetic reference images.
Another study, Yarnykh, Vasily L., Time-Efficient, High-Resolution, Whole Brain Three-Dimensional Macromolecular Proton Fraction Mapping, Magnetic Resonance in Medicine (2015), which is which is incorporated herein by reference in its entirety, discusses results obtained using systems and methods such as those described with reference to FIGS. 1-6. MPF maps were obtained based on data from 8 healthy subjects (4 females, 4 males, age range 29-66 years). Additionally, standard anatomical high-resolution 3D magnetization prepared rapid acquisition gradient echo (MPRAGE) T.sub.1 weighted images (TR/TE=5.9 ms/2.8 ms, FA=8.degree., inversion time (TI)=900 ms, shot interval=2500 ms, voxel size=1.times.1.times.1 mm.sup.3, scan time=4 min 44 s) were acquired with the same field of view for comparison purposes.
MPF maps were reconstructed with the standard method based on the acquired reference image and the synthetic reference method described above. Previously determined constraints for non-adjustable two-pool model parameters and their combinations were used in both algorithms as follows: cross-relaxation rate constant R=19.0 s.sup.-1, product of R.sub.1 and T.sub.2 of free water protons R.sub.1 T.sub.2.sup.F=0.022, and T.sub.2 of bound macromolecular protons T.sub.2.sup.B=10 .mu.s.
Before reconstruction, non-brain tissues were removed from source images by applying a brain mask created from the PD-weighted image. Additionally, CSF was segmented out from R.sub.1 maps based on a threshold value of R.sub.1=0.33 s.sup.-1 (equivalent to T.sub.1=3000 ms). MPF maps reconstructed with both methods were segmented into WM and GM. Segmentation was performed by an automated segmentation tool using software in the native image space with a Markov random field weighting parameter 0.25. To account for potentially incomplete CSF segmentation by R.sub.1 threshold and exclude voxels containing partial volume of CSF (PVCSF), a third mixed tissue class was also prescribed. Tissue classes were defined by specifying initial tissue-type priors with the following MPF values: 12% for WM, 6% for GM, and 1% for PVCSF. Resulting binary segmentation masks were used to calculate mean MPF in WM and GM.
A statistical analysis was carried out in SPSS software. Normality of data distribution was checked using the Shapiro-Wilk test to justify parametric analyses. An agreement between MPF map reconstruction techniques with acquired and synthetic reference images was assessed using Bland-Altman plots. The bias between MPF values in segmented WM and GM measured from MPF maps reconstructed by the two methods was examined using the one sample t-test for the mean difference between paired measurements. The limits of agreement were calculated as the mean difference .+-.1.96 standard deviation of the mean difference. To estimate variability between measurements from the two reconstruction options, the within-subject coefficients of variation (CV) was calculated as the percentage ratio of standard deviation to the mean of paired measurements. The Levene's test was used to assess equality of variances. P values less than 0.05 were considered statistically significant.
FIG. 7 depicts a series of images 700 obtained from a system such as the system 100 of FIG. 1, in accordance with at least one embodiment. The images were obtained from the study discussed above and depict the minimal set of three source images 710 (representing T.sub.1-weighted (T.sub.1W), PD-weighted (PDW), and MT-weighted (MTW) GRE images), B.sub.0 and B.sub.1 field maps 720, and corresponding parametric maps 730 that were generated by the algorithm based on the source images and field maps. The parametric maps 730 include R.sub.1 and PD maps reconstructed from T.sub.1W and PDW images, a synthetic reference image, a resulting MPF map, and a binary segmentation mask corresponding to WM, GM, and PVCSF tissue classes. The images 700 are presented with removed extra-cranial tissues and CSF.
FIG. 8 depicts a comparison table 800 comparing acquired and synthetic images and MPF maps reconstructed with the acquired and synthetic images, in accordance with at least one embodiment. In the comparison table 800, acquired images, in column 810, and synthetic reference images, in column 820, are present in the top row of the table, and MPF maps reconstructed with the associated images are depicted below for each respective column. Column 830 of the comparison table 800 shows subtraction images to demonstrate the difference between the acquired and synthetic images and maps reconstructed from each. Only minor discrepancies associated with sub-voxel misregistration are seen on the tissue borders in column 830. Grayscale ranges correspond to MPF of 0-18% for the maps and -9-9% for their difference.
FIG. 9 depicts a graph 900 illustrating percent (MPF.sub.acq-MPF.sub.synt) plotted over percent (MPF.sub.acq+MPF.sub.synt)/2, in accordance with at least one embodiment. MPF in WM (black symbols) and GM (gray symbols) demonstrated close agreement with no significant bias between reconstruction methods, equal variances, and small within-subject coefficients of variation. Solid and dashed lines correspond to the mean difference and limits of agreement, respectively.
FIG. 10 depicts a table 1000 illustrating a comparison between MPF measurements with acquired and synthetic reference images in segmented brain tissues in accordance with at least one exemplary embodiment. Results of quantitative MPF measurements in segmented WM and GM based on the maps reconstructed with acquired and synthetic reference images are provided in the table 1000.
FIG. 11 depicts reformatted sections 1100 of a 3D MPF map, R.sub.1 map, and T.sub.1-weighted image comparing tissue contrast for iron-rich GM anatomical structures. Arrows indicate the globus pallidus (GP), substantia nigra (SN), and dentate nucleus (DN). Arrowheads indicate intrinsic variations of MPF within WM. As shown in FIG. 11, the MPF maps can produce pure GM contrast for subcortical structures with high iron content, such as GP, SN, and DN. These structures appear clearly hypointense on MPF maps compared to WM, similar to the rest of GM. In contrast, iron-rich GM structures are undistinguishable or poorly distinguishable from surrounding WM on both T.sub.1-weighted anatomical images and R.sub.1 maps. FIG. 11 also shows the contrast variations within WM due to a variable degree of myelination. Such an inherent WM signal variability is much less clear or invisible on T.sub.1-weighted anatomical images and R.sub.1 maps.
In yet another study, Yarnykh V. L., Three-Dimensional Macromolecular Proton Fraction Mapping of the Human Cervical Spinal Cord, Proceedings of the 22th Annual Meeting of ISMRM-ESMRMB, Milan, Italy, 2014; p. 3449, hereby incorporated by reference in its entirety, systems and methods such as those discussed above with reference to FIGS. 1-2 are evaluated. In this study, single-point MPF mapping was combined with two-point variable flip-angle (VFA) T.sub.1 mapping. High signal-to-noise ratio (SNR) is important for the accuracy of single-point MPF mapping and the use of source images with SNR>100 has been recommended. To gain SNR, the multiple gradient echo summation technique was applied. To reduce scan time, single-point MPF mapping was modified to avoid acquisition of a reference image by using a synthetic reference image.
The pulsed magnetization transfer matrix equation was iteratively solved by the Gauss-Newton method and standardized constraints for non-adjustable two-pool model parameters, as well as corrections for B.sub.0 and B.sub.1 inhomogeneities, to yield an MPF map. As input data, a single-point MPF measurement requires a GRE MT-weighted image, a reference GRE image for data normalization, and an independently acquired T.sub.1 map. The two-point VFA T.sub.1 mapping method utilizes a T.sub.1-weighted image due to the absence of a long off-resonance saturation pulse. Instead of acquiring a reference image with TR and flip angle (FA) corresponding to those for the MT-weighted image, it can be synthesized from T.sub.1 and PD maps reconstructed from VFA data according to the Ernst equation.
Data was obtained from 4 healthy subjects (2 females, 2 males, age 39-52 years). Images of the cervical spinal cord were acquired on a 3T scanner with a manufacturer's phased array spine coil. PD and T.sub.1-weighted GRE images were acquired with TR=20 ms and FA=3.degree. and 20.degree., respectively. MT-weighted GRE images were acquired with TR=40 ms, FA=10.degree., and off-resonance saturation pulse with single-lobe sinc-gauss shape, offset frequency 4 kHz, effective saturation FA 560.degree., and a duration of 18 ms. All images were acquired in the 3D mode with FOV=140.times.140.times.160 mm.sup.3, and voxel size of 1.times.1.times.2 mm.sup.3. Four-echo acquisition acquisition was used in the sequences. Dual-echo B.sub.0 maps and actual flip-angle imaging (AFI) B.sub.1 maps were acquired with voxel size of 1.5.times.1.5.times.4 mm.sup.3. Scan times were 3 min 45 s for the T.sub.1 and PD-weighted images, 7 min 30 s for the MT-weighted image, 2 min 25 s for the B.sub.0 map, and 4 min 45 s for the B.sub.1 map, thus resulting in about a 22 min acquisition time for the entire protocol.
To evaluate performance of multiple echo summation, MPF map reconstructions were carried out with the first echo image only and sums of 2, 3, and 4 echoes. SNRs in the spinal cord relative to the background noise were estimated on MT-weighted images at the central location of the slab. MPF values in anatomical structures of the spinal cord were measured on the MPF maps. Measurements were performed in WM regions of interest corresponding to ventral, lateral, and dorsal columns and GM regions of interest corresponding to ventral and dorsal horns. A series of paired t-tests on bilaterally averaged data pooled across subjects and levels was used to assess regional differences between anatomical structures.
FIG. 12 depicts an overview of the spinal cord MPF mapping 1200, in accordance with at least one exemplary embodiment. FIG. 12 shows component images involved in data acquisition and MPF map reconstruction. The top row 1210 shows acquired source GRE images: PD-weighted (PDWI), T.sub.1-weighted (T.sub.1WI), and MT-weighted (MTWI). The middle row 1220 shows intermediate parametric images reconstructed from VFA data (PD map, T.sub.1 map, and a synthetic reference image). The bottom row 1230 shows B.sub.0 and B.sub.1 field maps and a resulting MPF map. A dramatic increase of the contrast between WM and GM in the spinal cord compared to source images or other parametric maps (T.sub.1 and PD). Multiple echo summation allowed about 1.4, 1.6, and 1.8-fold SNR gain in source images for 2, 3, and 4 echoes, respectively.
FIG. 13 depicts images 1300 of the effect of multiple echo summation on the SNR, MT-weighted contrast, and quality of MPF maps, in accordance with at least one exemplary embodiment. In FIG. 13, the use of multiple echo summation visibly improves MT-weighted contrast on source images and anatomical consistency of spinal cord structures on reconstructed MPF maps. A top row 1310 shows MT-weighted images corresponding to the first-echo (N.sub.echo=1) and summation of 2, 3, and 4 echoes (N.sub.echo=2, 3, and 4) with SNR values in the spinal cord. A bottom row 1320 shows MPF maps reconstructed from the images in the top row 1310.
FIG. 14 depicts images 1400 of anatomic appearance of WM and GM on MPF maps of the cervical spinal cord at different levels, in accordance with at least one exemplary embodiment. In FIG. 14, representative axial sections of the 3D MPF map of the cervical spinal cord at the levels of C2, C3, C4, and C5 vertebra were obtained from a healthy subject. The image 1410 depicts a reformatted middle sagittal section of the same 3D MPF map.
This study demonstrates feasibility of 3D MPF mapping of the human spinal cord in vivo. By using the minimal possible dataset comprising only three source images and multiple echo summation, the described approach enables sufficient resolution and SNR with a reasonable scan time. Quantitative myelin assessment and high-contrast neuroanatomical imaging are combined for a variety of spinal cord applications.
Besides quantitative tissue characterization, high-resolution MPF mapping has a potential to become a useful technique for various neuroscience applications including segmentation of cortical, subcortical, cerebellar, and brain stem structures. The use of MPF maps as source images in such applications may be advantageous due to a very high tissue contrast and insensitivity to iron. Another important advantage of MPF maps as source images is the inherent correction of field and coil reception non-uniformities. Specifically, spatial variations of B.sub.0 and B.sub.1 fields are corrected based on the rigorous physical model by directly accommodating corresponding field maps into the image reconstruction workflow. The effect of the coil reception profile is eliminated by normalization to the synthetic reference image and does not propagate into resulting MPF maps. As such, the use of MPF maps eliminates the need in computationally-intense bias field correction algorithms and allows utilization of native quantitative voxel values in various post-processing tasks.
In yet another study, first reported herein, systems and methods such as those discussed above with reference to FIGS. 1-3 are evaluated. In this study, a feasibility of measuring age-related changes in brain myelination in children using non-localized whole-head MPF measurements was assessed. As a model of a non-localized magnetic resonance signal measurement, a whole-head three dimensional image was used, and intensities of all voxels in such an image were averaged to yield a single value proportional to a magnitude signal that could be obtained from an entire head without applying magnetic field gradients. Data were acquired using three radiofrequency modes as depicted in FIG. 3 and processed using equations -, as specified above. A clinical MRI scanner (Toshiba.RTM. Vantage Titan) with 1.5T magnetic field strength was used for data acquisition. The three radiofrequency modes depicted in FIG. 2 were executed with following parameters: .alpha..sub.1=3.degree. and TR=16 ms for the first mode; .alpha..sub.2=18.degree. and TR=16 ms for the second mode; and .alpha..sub.3=8.degree., TR=20.5 ms, .DELTA..sub.mt=2 kHz, .alpha..sub.mt=200.degree., and t.sub.mt=1.5 ms for the third mode. 3D images corresponding to each mode were obtained with a large field-of-view covering the whole head and isotropic spatial resolution of 1.4 mm.sup.3. Data were obtained from 23 children aged from one week to 14 years. All participants underwent MRI for various clinical indications but were free from radiological abnormalities of the brain according to evaluation by certified neuroradiologists.
FIG. 15 depicts example cross-sections 1500 of color-coded 3D MPF maps obtained from subjects of one month and of one year of age. A dramatic increase of MPF caused by the development of myelin in white matter is easily visible in FIG. 15 by comparing MPF maps from the subjects of different age.
FIG. 16 depicts a graph 1600 illustrating a dependence of the whole-head MPF values obtained from average intensities of all voxels in 3D images on the age for 23 children, wherein each point represents a whole-head MPF measurement for a single study participant. A solid line in FIG. 16 depicts a plot of a generalized logistic model fitted to experimental data. FIG. 16 illustrates that dramatic changes in brain myelination over the first year of life can be reliably captured by non-localized whole-head MPF measurements.
FIG. 17 depicts a graph 1700 of Pearson's correlation between whole-head MPF measurements and age after a logarithmic transformation of the data presented FIG. 16. The correlation coefficient is high (r=0.987), thus indicating that whole-head MPF values describe the process of brain myelination with high quantitative accuracy.
This study demonstrates that the MPF measured from the whole-head magnetic resonance signal is representative of the whole-brain MPF and provides a realistic non-invasive measure of the brain myelination process during the childhood. Even without suppression of unwanted signals from tissues not belonging to the brain, a prototype non-localized method for MPF measurements, such as that depicted in FIG. 3, accurately characterizes quantitative age-related changes in the brain myelin content. Further implementation of an advanced method with suppression of unwanted tissue signals, as depicted in FIG. 4, is expected to improve sensitivity of non-localized MPF measurements to more subtle changes in myelination, particularly to those associated with a delayed development in children and neurological disease in adults.
While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims, along with the full scope of equivalents to which such claims are entitled. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting.
| 2019-04-20T00:39:02 |
http://patents.com/us-20160120456.html
|
0.999998 |
What: "Logic is like a laser - a tool whose best use is not illumination, but rather focus. A laser may not provide light for your home, but, like logic, its great power resides in its precision. The following sections describe just a few areas in which logic is commonly used." The link has a list of pursuits that use loigic.
What: "In comparison to your heart, logic is a flimflam. Logic is like a little martinet who has a schedule all laid out and wants it to be followed because it's a system and its logic is iron-tight." A lofty discussion of the heart and logic and how they don't mix, at least in the author's mind.
What: "Logic is like a right angle. It is not organic. It is something that we create. It is like a perfect circle. When Kepler searched relentlessly for proof that the universe moved according to the five Platonic solids, he was mistaking outer space for inner space."
What: "Logic is like Sausages." is the comment, and the response is "It's soft and chewy, springy, distasteful yet satisfying." The link is a discussion in mental chunks of stuffs, as I like to say.
The ladder of logic: What's on the bottom?
What: "Logic is like a ladder - by means of it you go up step by step. But how are you going to prove that the bottom of it is all right? The ladder rests on the ground; but what supports the ground?" This is the fundamental question of this time, and most other times. What's on the bottom of your ladder?
What: 'Logic is like a river that can be followed from a source (a set of premises) to a set of conclusions consistently deducible from those premises. The problem is, there are many forces in our ape brain that try to distort the flow of that river. These distorting influences include emotions, ego, agendas, and preconceived representational categories." See the link to read the whole argument.
What: "But logic suffers from an inherent weakness: it can be used as powerfully both for and against a proposition. If one wants to describe logic rightly, and those who know it well have said so, logic is like a lawyer or a prostitute who goes with anyone who pays the fee." A discussion of using logic to justify or prove that reincarnation occurs.
"Contemplating geologic time is like doing brain yoga. It's a stretch to think in millions of years, so I start slow. After all, continents drift leisurely."
"The key word here is "also impending". They say Hekla will erupt in the near future which can be interpreted pretty much as anything. Can erupt tomorrow or in a years time. Geologic time is like our coffebreaks...Alot longer than you might think."
"Dr. Blauvelt pointed to an analogy to help explain why generic drugs and biosimilars, after all is said and done, may not be analogous products after all: "If aspirin is like a bike and if human growth hormone is like a car, then a biologic is like a jet plane. It might be relatively simple to copy a bike, but remaking a jet plane is ..."
| 2019-04-21T11:19:33 |
http://www.metamia.com/analogize.php?q=logic
|
0.999999 |
Does the encoder implement automatic bandwidth limitations, lower sampling rates, at low bit rates?
Hi. Yes, at quality settings lower than 5 (which is --standard), the maximum bandwidth is lower than 20 kHz.
At --quality 7 it's 22 kHz, --quality 5: 20 kHz, --quality 4.5: 17.9 kHz, --quality 3: 13.1 kHz, etc.
What is the bandwidth on the lowest setting?
The lowest setting is --quality 0.1 and its maximum bandwidth is 5.5 kHz.
| 2019-04-19T16:26:36 |
https://forum.musepack.net/showthread.php?s=9ab23510af105b0b776642915256f321&p=2700
|
0.999999 |
Healthline’s core business – providing health information to 70MM monthly users - has scaled well for 2+ years and the business was ready to expand. But expand to what exactly? DONECOMPANY helped move Healthline quickly from business ideas to tangible high-revenue market opportunities.
Bringing any new product to market is high risk and high cost. It's not the time for guessing or gut calls. Our job is to mitigate those risks by using data and design to increase the understanding of the opportunity.
Mine available data: After a deep dive into Healthline’s user analytics, we outlined the users, experiences, and differentiators that best aligned with their overall mission.
Research & analyze: We used quantitative research and further data-analysis to identify 9 key health motivations that signified business opportunities for a tightly defined consumer.
To get from “could be” to a solid new opportunity, it's critical to evaluate the risks, failure points, monetization, capabilities, and go-to-market pitfalls that could indicate success or failure.
Generate ideas: Within a week, we moved Healthline from structured brainstorming on a broad swath of ideas to a rigorous evaluation process. The end result was 9 new product opportunities that had broad appeal, addressed major audience motivations, and had significant monetization upside.
Immediate User Input: We use quantitative surveys to identify the relative strength of new service ideas. As a result we identified 4 new services with an "interest" score over 80%; and 3 with a "likelihood to use" score over 70%.
To gain the most confidence in new innovations and reduce risk, we design prototypes to maximize learning. We produced over 12 UX models and 3 clickable prototypes - including a live-store test - and research with target users.
Prototype tests: We identified the best position for market differentiation, highest value features and promises, and possible failure points.
Handoff: The Healthline team could take over the prototypes and with direct guidance on product features, go-to-market strategy, design and start experiences.
Our goal for innovating new products with clients is to hand-off a potential new product where solid demand exists and an understanding of how that demand can best be exploited. We succeeded in uncovering break-though products for Healthline that achieve ambitious goals and have user validation within 10 weeks.
| 2019-04-21T12:30:40 |
http://www.done-co.com/projects/healthline/prototyping
|
0.998595 |
Over the past few years, debate has grown over whether we need to raise our nation's minimum wage and the wages of fast food and other low-wage workers. These workers and their allies have pressured companies to raise salaries through strikes, protests, and campaigns. Very recently, McDonald's and other fast food and retail employers have raised their hourly rates as a result of this outcry. But many say these increases are too small and only impact a small percentage of the total workforce (in the case of McDonald's).
Should Fast Food Workers Get a Raise?
Note: Before the lesson, reflect on whether you have students in your class whose parents, or any students themselves, who work in the fast food industry or would be considered a "low wage worker." Students sometimes feel relieved to discuss a topic that is relevant to their lives, while others may feel awkward or embarrassed. This should not preclude you discussing the topic; however, be careful not to ask or call attention to students in the class who relate to the topic on a personal level. Consider talking with students in advance if you feel it may be a sensitive subject for them.
Students will reflect on their perceptions of the characteristics of fast food workers.
Students will learn more about the characteristics of fast food workers, minimum wage and the current movement to increase minimum wage and salaries for fast food and other low-wage employees.
Students will be able to calculate the theoretical income and expenses for a low-wage employee and assess whether it is a "living wage."
Students will explore their own perspectives about minimum wage and low-wage employees' salaries and represent that position in an argumentative essay.
Who are the fast food workers?
Explain to students that you are going to ask them to reflect on who they think are the majority of fast food employees, including some identity characteristics or demographics of that group. Have students take out a piece of paper and pen and explain that they will take notes as you read questions aloud. Tell students that for each of the questions, they should jot down their first thoughts.
What do you think is the average age of a fast food employee or what is the age of most fast food workers?
What race do you think most fast food workers are?
Do you think fast food workers are mostly male or female? How far in their formal education did they go?
What percentage do you think have been to college? How many hours do most fast food employees work?
What is their average wage per hour?
After responding to the questions, have students turn and talk to a person sitting next to them and compare their notes, taking a minute or two to share both ways. Then have students come back to the larger group and ask a few students to share aloud what they wrote down.
In what ways were your ideas about fast food workers similar to or different from what you see on the chart?
Where do you think you received your ideas about the characteristics of fast food workers?
Are these stereotypes and/or misconceptions? Why or why not?
Do you think most people would be surprised or not surprised to learn this information?
Given the information you now have, do you think there are any special considerations for people who work in the fast food industry?
Ask students: Have you heard anything about McDonald's in the news recently? If they haven't heard, explain that recently, McDonald's announced that they are offering raises of $1 an hour to their employees. Tell students that we will discuss more about what prompted this raise later in the lesson.
Ask: What is minimum wage? If they don't know, ask them to figure it out by defining minimum (the least quantity or amount possible) and wage (money that is paid or received for work/services). Come to a definition of minimum wage as: the lowest wage payable to employees in general or as fixed by law or union agreement. Explain that there is a nationwide federal minimum wage, which is now $7.25 an hour and each state has its own minimum wage. Use the Minimum Wage In Your State chart (or Minimum Wage Rates for 2015) to help students find the minimum wage for their state. In addition, in some states the minimum wage varies from one location to another within the state.
Ask students: What is a "living wage"? Explain that a living wage is the income needed to meet the minimum standards of living based on the region of the country and type of community (urban, suburban, rural, etc.) in which one lives. For more information about living wage, see Living Wage Calculator.
Explain that over the past few years, there has been a lot of public conversation and activism around increasing the minimum wage and specifically, increasing wages for fast food workers and other low-wage employees. Retail stores such as Target, Walmart, TJ Maxx, etc., have recently pledged to increase their hourly wage. Employees say that their low salaries are not enough to live on (i.e. a "living wage") and they often have to get a second job or receive food stamps and other forms of public assistance to compensate for the low wages.
Explain that companies that don't support an increased minimum wage or $15/hour argue that increasing labor costs will force them to increase the price of their products, which will be burden for customers - or drive customers away. Some small business owners say that requiring them to pay a higher minimum wage will force them to hire fewer employees or even shut down.
To help students learn more, have them read the following article: McDonald's Workers, Vowing a Fight, Say Raises Are Too Little for Too Few. Give students 10 minutes to silently read the article.
Can you tell what the author's point of view is?
How many McDonald's employees got a raise? What is the approximate percentage of the total employees who received an increase in their salary?
Why are workers not happy about the raise?
How would you feel if you were in their situation?
Math Activity: is minimum wage a living wage?
With students, determine the salary of working at McDonald's: Use the average McDonald's salary ($8.94 per hour) and the average number of hours worked (24 hours per week) to calculate the average income.
If you don't have time or accurate information for a local and specific calculation, you can share this The Real Budgets Of McDonald's Workers with students and have them reflect on the income and expenses of the individuals they presented.
Do you think a family could live on a salary from McDonald's or other fast food chains? Why or why not?
What do you think fast food companies should do about this?
Do you think it's important that fast food companies provide a living wage for its employees? Why or why not?
Explain to students that now they will have an opportunity to explore their own opinions and learn about their classmates' points of view about minimum wage and fast food and other low-wage employee salaries.
Use the following five statements and post them around the room, either using chart paper (one statement per paper) or use the Opinion Statements handouts provided. Distribute different color post-its, dots or markers for students to distinguish their point of view (e.g. green=agree; red=disagree and yellow= not sure or mixed feelings).
Fast food workers do not need a raise.
The federal minimum wage should be at least $15 an hour.
If we raise the minimum wage, companies will have less jobs to offer.
Fast food and other low-wage workers should be able to work 40 hours per week if they want to.
It is not these companies' responsibility to provide a living wage for their employees.
Instruct students to move around the room to each of the five statements and using post-its, dots or markers (drawing a dot), signify what they think on each of the five statements, either agreeing (green), disagreeing (red) or not sure/mixed feelings (yellow). Give students 5-7 minutes for this task.
Was it difficult to decide what your opinion was on any of the statements? Which ones and why?
What do you notice about the charts?
Are there are any patterns that you notice?
Did you change your position from what you might have thought before the lesson?
Would anyone like to share where they placed their mark and why?
Have students reflect on the statements and the opinions discussed above. Explain that they will now write an argumentative essay using one of the statements as a basis for their paper. They should choose the statement they feel most strongly about. This assignment can begin in class but then should be assigned for homework; it will take at least several days since it involves research. Explain to students that they are to write an argumentative essay that argues a position and uses research, quotes, examples and data to convey their position well.
Gather evidence such as examples, statistics, quotes, polls, facts, etc.
o 2-3 reasons for your opinion with supporting evidence (examples, historical information, statistics, polls, quote and facts).
As a culmination, have students present their papers in class and consider publishing them online or submit them to a local paper for the op-ed or letter to the editor sections.
In this short activity, pegged to a planned nationwide strike by low-wage workers on April 15, students discuss the growing movement to raise the minimum wage to $15 an hour.
The nation's largest private employer just raised its wages. This short activity has students consider the significance of Walmart's move and what motivated it.
This lesson focuses on a less well-known part of Martin Luther King Jr.’s legacy, but one that is extremely relevant today: the Poor People’s Campaign that Dr. King led over 40 years ago. The lesson links this campaign to current struggles to combat poverty in the US, including efforts by workers at fast food restaurants, Wal-Mart, and others to substantially increase their wages and those of millions of other Americans by raising the federal minimum wage.
| 2019-04-21T12:44:33 |
https://www.morningsidecenter.org/teachable-moment/lessons/exploration-raise-low-wage-workers
|
0.787955 |
Despite today's low gas storage, the market is oversupplied.
The market consensus has priced in continued oversupply going forward, but early evidence suggests a meaningful production slowdown in 2019.
As slow supply growth undershoots consensus expectations, I see high odds of a tighter market and higher prices in the second half of 2019.
Today, the natural gas (UNG) market is oversupplied and prices remain stubbornly around $3 - about the same price as 6 months ago. However, I still believe prices could soar as we move through 2019. The reason? Production will fall short of market expectations.
In this article, I'll explain why natural gas production growth will slow dramatically in 2019. This slowing production will meet the new demand growth in the second half of the year, tightening the market and sending prices much higher. But first, let's see where the market stands today.
On the surface, this storage deficit might seem to indicate an undersupplied market. But today's deficit is purely a function of a low starting point. You see, we entered last winter with the lowest natural gas inventory in 13 years. This low starting point combined with an early-November cold blast to ignite an epic short squeeze, fueled by offsides traders - like the now-infamous OptionsSellers.com.
(And don't let the weepy-eyed apology video fool you. Anyone with even a cursory knowledge of derivatives knows the immense risk of selling naked commodities options on leverage. OptionsSellers.com did not suffer from a "rogue wave." They could have easily hedged their risk by selling call spreads instead of naked calls. But that would have given up their potential upside. So they sold naked call options for maximum gains and thus exposed their clients to catastrophic risk. That's just greed, plain and simple. And now I'll jump off the soapbox).
After November's brief short-covering rally, one of the warmest Decembers in years was followed by an unseasonably warm January, sending gas demand and prices tumbling back down. And despite gas inventories plumbing 4-year lows, prices remain stubbornly below $3 on both a spot basis and through much of the 2019 - 2020 forward curve. That means traders are convinced that an oversupplied gas market will erase today's gas deficit during the upcoming storage refill season (April 1st through October 31st).
First, let's compare the coldness of this past winter (2018/2019) versus the previous year (2017/2018) using "heating degree day" (HDD) data. Heating degree days simply measure how cold a given region is versus a baseline temperature. In the U.S., the standard baseline is 65 °F. So a 35°F day would generate 30 HDDs, and two 35°F days would generate 60 HDDs, and so on.
As a very rough approximation, this winter's 5% greater heating demand and 17% lower storage draws implies we were oversupplied by about 475 billion cubic feet during this past winter, or roughly 3 Bcf/d. (Author's Note: Don't take this as gospel; it's a very rough approximation for the sake of argument.) Now, 3 Bcf/d might not sound like much on the surface, but consider this….
The U.S. ended the winter gas drawdown season with 1,107 Bcf of gas in storage on March 29th. That's 505 Bcf below the 5-year average end-of-winter storage level of 1,635 Bcf. And with the storage refill season spanning just over 200 days, you can see how a persistent 3 Bcf/d surplus could generate over 600 Bcf of excess supply. That would transform today's 500 Bcf deficit into a mild surplus by the start of next winter. Now that's just a hypothetical example, but it happens to fall within the ballpark range of the official projection from the Energy Information Agency ("EIA").
In the EIA's latest short-term energy outlook, the agency projects that we will almost completely erase today's storage deficit and end the refill season with gas storage just 1% below the 5-year average. Said differently, the EIA projects that in the 213 days from April through October, the market will supply an excess of roughly 500 Bcf - or 2.3 Bcf/d. And with nearly every gas futures contract in 2019 trading under $3, the market consensus has priced in a similar forward outlook.
Do you believe the EIA production forecast?
I recently explained how investor demands for less growth and more cash generation were forcing producer restraint in the shale patch. That trend is showing up in capex cuts across every major shale gas producer I track. The list includes EQT (EQT), Range Resources (RRC), CNX Resources (CNX), Antero Resources (AR), Southwestern Energy (SWN), Cabot Oil and Gas (COG) and Gulfport Energy (GPOR).
If this trend persists, there's every reason to believe that Appalachian production growth could slow dramatically in the region. And this drop in drilling permits becomes even more bullish when you consider DUC (drilled but uncompleted) well dynamics.
Last year alone, Appalachian drillers burned through nearly 200 DUCs to deliver 2018's record production spurt. And with only about 500 DUCs left in Appalachian inventory, can they repeat this feat in 2019? Unlikely.
Remember, over the last few months, investors have clamored for shale companies to start generating free cash flow. So there's never been a greater incentive for shale drillers to tap DUC inventories and boost cash flow with minimal capex. And yet, the rate of DUC drawdowns has slowed to a crawl. Occam's razor says Appalachian shale drillers have stopped tapping their DUCs for one simple reason - they've run out. This is just my hunch, but time will tell.
The bottom line: less capex plus lower DUC inventories equals a big drop in Appalachian wells coming online in 2019. Less wells mean a lower rate of production. And the early production data so far confirms this view.
Has U.S. Gas Production Stalled?
This January, U.S. gas production fell to 88.6 Bcf/d, down from 88.9 Bcf/d in December. Admittedly, that's a small decline. But remember, we don't need an outright decline to balance the market. Anything less than blistering growth is a big positive for balancing the market, and a small decline is even better. That's because of the huge new demand coming online from ramping U.S LNG export capacity, plus pipeline exports to Mexico and gas-fired power demand.
Of course, these estimates could get revised higher in the coming weeks. But for now, it's simply one more piece of the production puzzle indicating a meaningful slowdown in the rate of growth this year versus 2018. As we move into the second half of the year, we start lapping last year's massive production increase. That means shale drillers will need to step on the gas to hit the EIA's 7.6 Bcf/d production target in the second half of the year. But based on the depleted DUC inventories, lowered capex, and collapse in Pennsylvania drilling permits - I see low odds of that scenario unfolding. That's why I remain bullish on natural gas as we move into the second half of 2019.
Disclosure: I am/we are long RRC. I wrote this article myself, and it expresses my own opinions. I am not receiving compensation for it (other than from Seeking Alpha). I have no business relationship with any company whose stock is mentioned in this article.
| 2019-04-23T18:33:00 |
https://seekingalpha.com/article/4254224-natural-gas-oversupplied-prices-still-double-2019
|
0.999711 |
"Millions of families can barely breathe," Warren said, in a feisty speech that struck aggressively populist and unapologetically left-leaning notes.
"It is not right." The Massachusetts senator -- who had announced her intention to run on New Year's Eve -- is among the highest-profile of the growing pool of Democrats hoping to unseat President Donald Trump in 2020.
The Democratic field is already the party's most diverse ever -- in gender, age and ethnic background -- and one of its more progressive. It includes several well-known women lawmakers, with Senator Amy Klobuchar expected to join their numbers on Sunday.
Warren's past battles with Wall Street have brought her a large following, and her campaign team has drawn the grudging respect of its rivals. Hoping to ride the momentum of her Lawrence speech, she heads next to early-voting Iowa and New Hampshire, followed by five other states.
But for now it is unclear how badly damaged Warren is by the stubborn controversy over her claim to Native American roots -- a claim Trump has seized upon to belittle her, mocking her as "Pocahontas."
'Exposed as a fraud': Trump's re-election campaign issued a dismissive statement ahead of Warren's announcement, saying she had "been exposed as a fraud by the Native Americans she impersonated and disrespected to advance her professional career." It said her "socialist ideas" would hurt workers.
Lawrence, the scene of Warren's announcement, is a former mill town where a group of women workers, including many immigrants, launched a strike in 1912 that spread across the region and came to be seen as a historic victory for women and for labor, with improved wages and working conditions.
She called for "big, structural change" in America that would reach beyond new US leadership -- though she called the Trump administration "the most corrupt in living memory."
Warren tried in her speech to appeal to a broad ethnic coalition, saying, "We must not allow those with power to weaponize hatred and bigotry to divide us."
Warren recounted her decision to go to law school when Amelia was not quite two years old, but said the only day care she could find -- with five days before classes were to start -- would take the child only if potty-trained.
A determined Warren accomplished the task, she told a laughing crowd, "courtesy of three bags of M&Ms." "Since that day, I've never let anyone tell me that anything is 'too hard,'" she quipped.
| 2019-04-26T11:51:52 |
https://smartinvestor.business-standard.com/market/Marketnews-569847-storydet-Democrat_Elizabeth_Warren_officially_kicks_off_her_2020_White_House_run.htm
|
0.999857 |
What are the different "Chapters" of benefits under the GI Bill?
Chapter 30-Montgomery GI Bill Chapter 30 benefits affect veterans with active duty service in any branch of the armed services of the United States. There are two different levels of benefits under Chapter 30. One level is for veterans who began active duty after 06/30/1985. This is by far the largest group of veterans.
The second level is for veterans who were eligible to receive benefits under the old Chapter 34 (Veteran's Education Assistance Program, or VEAP) on 12/31/1989.
The scale below is ONLY for the first level of veterans: those who began active duty service after 06/30/1985.
* 1/2 time or less: pays the amount of tuition and fees, not to exceed the amounts listed.
-Is in need of vocational rehabilitation because his or her disability creates an employment handicap.
Benefits under Ch. 31 can be provided for up to 48 months, and the veteran generally has 12 years from the time they are notified of entitlement to VA compensation to use their benefits. Each veteran will be evaluated by a case worker for interests, aptitudes, education, work experience, and vocational abilities. Benefits are paid directly to the school for tuition, fees and books and the veteran will receive a living stipend.
-You are a child (between the ages of 18 and 26, with some exceptions) of a veteran who is permanently and totally disabled due to a service-related condition; or who died in service; or who died of a service-connected disability; or who died while being evaluated as having total and permanent service -connected disability; or who is listed as a POW or MIA.
-You are the surviving spouse of a veteran who died of a service-connected disability, or died in service, or died while evaluated as having total and permanent disability resulting from a service-connected disability.
-You are the spouse of a veteran or service person who has a total and permanent disability resulting from a service-connected disability; or who is listed as a POW or MIA.
* Less than half-time will be given amount of tuition and fees only. Payment cannot exceed the listed amount.
Chapter 1606-Selected Reserve Chapter 1606 is the educational benefit program for active members of the Selected Reserve. Selected Reserve components include the Army Reserve, Naval Reserve, Air Force Reserve, Marine Corps Reserve, Coast Guard Reserve, Army National Guard, and Air National Guard. The Department of Defense and Department of Homeland Security determine who is eligible for Chapter 1606.
Like Chapter 30, Chapter 1606 recipients can recieve kickers as part of their enlistment contract. The amounts of the kickers available are $100, $200, and $350. You have to sign a kicker contract when you enlist in order to receive a kicker. An example of what a kicker can do: If you have a $350 kicker and you are a full-time student (which pays $375.00 per month as of 10/01/2017), then your monthly benefit will be $725.00 for each full month you are enrolled as a full-time student.
Chapter 1607-Reserve Educational Assistance Program (REAP)Chapter 1607 is for members of the Selected Reserve who have been called to active duty for at least 90 days since 09/11/2001. The Department of Defense determines eligibility for Chapter 1607 benefits based on the Selected Reserve member being called to active duty in a contingency operation declared by the President or Congress.
Chapter 1607 pays a percentage of the Chapter 30 rate for veterans with 3 or more years of service. For example, if you have at least 90 days but less than one year of active duty service, you receive 40% of the Chapter 30 rate, At least one year but less than two years pays 60%, and two or more years pays 80%.
-Have served at least 90 consecutive days of full-time National Guard duty for the purpose of responding to a national emergency which is supported by federal funds.
Chapter 1607 benefits are being phased out by the VA. In most cases, if a person was using Chapter 1607 benefits on November 24, 2015, they will be allowed to exhaust those benefits before November 25, 2019.
* - Below half-time pays Tuition and Fees only. Amount paid cannot be above amount listed.
Chapter 1607 recipients are also still eligible for "Kickers" that they signed up for when they joined the National Guard.
*NOTE – The housing allowance and books and supplies stipend are not payable to individuals on active duty. The housing allowance is not payable to those pursuing training at half time or less.
So, you will receive a PORTION of the 100% benefit based on your length of active duty service.
Important Note: Once you choose to get benefits under Chapter 33, you will not be able to choose a different chapter. It is an irrevocable decision.
Your Post-9/11 benefits can also be transferred to a spouse or dependent child, but you must usually make this transfer BEFORE you leave active-duty service. Other limits apply on transfer-ability, including how long you have to have been on active duty before you can transfer the benefits.
| 2019-04-19T02:53:03 |
http://www.noc.edu/chapters
|
0.991681 |
What makes a laser a laser?
In the most general sense I know, it is the high quality of the light. All of the light is going in the exact same direction, is in phase, has the same wavelength, and is concentrated into a small area.
| 2019-04-22T22:13:36 |
http://a-blast.org/blast/laser.html
|
0.999991 |
Understanding underlying mechanisms of cultural adaptations To generate new and generalizable scientific knowledge, it is beneficial to understand an intervention's mechanisms of change, as this understanding can identify critical intervention components that influence behavioral and other mediators that in turn influence the intended therapeutic outcomes ,. As they become more integrated into the new culture, they may even begin to start to question their identity within it. Orthogonal cultural identification theory: the cultural identification of minority adolescents. Those who were born at high altitude tend to develop larger lung capacities than do those who were not born at high altitude, but moved there later in life. This example illustrates how investigators may reframe their perspective on culture by addressing a common cultural identity, such as adjudicated youths, as the cultural unit of analysis for the design of a culturally relevant adaptation, in place of focusing on the label of ethnicity per se. This dynamic tension can foment controversy, raise contentious issues, and launch a search for common ground. Genetic Change When an environmental stress is constant and lasts for many generations, successful adaptation may develop through biological evolution.
Especially if a society is rather new or has little memory of old beliefs, the nevironment funds their entire culture. According to Arutyunov, we should regard culture as a set of different ways of institutionalizing human activity. During the past 40 years, anthropologists have discussed a broad spectrum of variation in parameters. We can trace the roots of this idea to physics in the 18th century. This is where they will overlook minor problems and look forward to learning new things. Also, as noted, the need for culturally adapted treatments might be lower for children than for adults. History of the Idea The origin of the concept of cultural adaptation and dissemination in contemporary anthropological literature is connected with the concept of cultural systems that, to a certain extent, fit the living conditions of their transmitters.
Mind you, they had better not try that with a large, fit Aussie woman or they might get more than they bargained for! A person in this position may encounter a wide array of emotions that the theory describes in four different stages. A dynamic, multi-level model of culture: from the micro level of the individual to the macro level of a global culture. Although the authors' confidence in this observation was attenuated by the small number of studies, they interpreted this pattern as suggestive evidence that low-acculturation participants are in greater need of a cultural adaptation and stand to benefit more. An adaptation is when an animalneeds new methods of survival. A proposed cultural adaptation should consider the influences of various cultural elements such as religion, gender, and social class ,. You not only have to learn about this new culture, but you must also be able to live and function within it.
It also competes well because a single plant can release thousands of seeds twice a year. Environmental psychologists have also adapted it. Ordering in a restaurant, having the Internet turned on in their apartment, or simply taking a taxi to the other side of town becomes more of an adventure. How do you cope with changing demands, uncertainty and stress?. The concepts and terminology used in planning an adaptation will influence this strategy.
Confusion can arise from gesture meanings or even the gender and ethnicity of the human figure. At the present time, the best answer to the question regarding the effectiveness of cultural adaptations has been provided in a meta-analytic review by. Another adaption that she made to try and fit in to our society was to drink alcohol which she had never done in her like as she had been brought up in a Muslim family! It is characterized by the feeling of culture shock, also described as the sensation of confusion associated with experiencing a new culture. Marketing The most effective way to adapt your marketing to local practices is to study the approaches of local companies. Language Language is an obvious but often misunderstood part of a country's culture. There, the cultural system no longer acts as an adaptive unit but only as a universal mechanism of adaptation.
What are some clear examples of cultural adaptation in humans?
Language is a distinct facet of culture that encodes symbols, meanings, forms of problem solving, and adaptations that also facilitate the group's survival , as cited by. Myth 2: When in Rome, act like the Romans This idea actually comes from ancient times — from the letters of St. The Center for Social Anthropology and Computing. As a result, western European and American science now thinks of cultural systems and societies as autonomous but mutually interdependent units in which complicated mechanisms of adaptation to living conditions are elaborated and realized. Or, what if you were trying to learn about this new culture, but people get frustrated when you do not understand right away? A related issue involves the cultural competence of the intervention developers and of the intervention delivery staff. In: Glanz M, Hartel C, editors.
Who we are does not only depend on our nature and personality but also the culture that surrounds us. An ecodevelopmental framework for organizing the influences on drug abuse: a developmental model of risk and protection. The process of being bicultural or intercultural usually begins with overcoming the language barrier. Someone in the adjustment stage would find this normal and be perfectly comfortable walking to the store. Within this duality, a dynamic tension has emerged between the standardized nomothetic scientific top-down approach that demands fidelity in its implementation and the idiographic casewise bottom-up approach that demands sensitivity and responsiveness to each person's unique needs. The evidence for the effectiveness of cultural adaptations is promising but mixed.
Biological adaptations vary in their length of time, anywhere from a few seconds for a reflex to a lifetime for developmental acclimatization or genetics. Such adaptive approaches hold considerable promise in general, and particularly for cultural adaptations that would be applied with specific and somewhat heterogeneous subcultural groups. Fundamentally, culture consists of the world-views and lifeways of a group of people. How might a cultural adaptation be effective for such a diverse population? Does manualization improve therapy outcomes? I know that even if I spoke in English, my intonation was noticeably Asian and I sometimes felt funny listening to myself. Associated Press Audio and video can also be culture specific.
The idea of optimum adaptive level is that the human group always tries to minimize the changes necessary for achieving an adaptive effect. In contrast, adaptive interventions are closer in form to individualized clinical practice because they provide explicit guidelines for the delivery of different dosages of intervention components depending on the unique needs of individual clients and based on evidence-based decision rules for determining variations in these dosages. They constantly move to areas having greener pastures and adequate water to ensure that their livestock are healthier; it also facilitates grass growing back again in areas that have been grazed by cattle. This model describes how events at each of these levels can influence an adolescent's identity development and behaviors. Accordingly, members of this group share a common group identity, although each individual also differs from the others in his or her unique combination of cultural tiles. As far as the written historical records of the country would show, the country has under been Spanish and American control thus explaining the Western tendencies.
| 2019-04-23T18:08:57 |
http://busanlottedfs.com/cultural-adaptation-example.html
|
0.999999 |
Halliburton company is listed under symbol HAL.
HAL is going down according our predicted trendThe price gives the following signals: We see a previous top at 2019-02-25 | 20, we see a previous top at 2019-02-15 | 20, we see a previous top at 2019-02-05 | 20, we see a previous top at 2019-01-25 | 20 and we see a previous top at 2019-01-18 | 20. Stochastic is going down | 30.
| 2019-04-20T00:17:57 |
https://stockpredictions.net/prediction/usa-stocks/16471/halliburton-company-hal
|
0.995944 |
Robert Reed's column ("State helps make racing a family affair at Arlington," May 15) hits the issue squarely when he says, ". . . there is ample evidence that gambling can become addictive and that children are particularly impressionable." Here is some evidence.
One of every four clients we treat at Haymarket Center for substance and/or alcohol abuse is also a compulsive gambler. If the gambling addiction is not addressed along with the primary addiction, 40% of those still gambling when they leave our programs will relapse into substance and/or alcohol abuse. Haymarket Center provides limited counseling to clients with co-existing gambling and substance abuse problems. Finally, casualties of gambling addiction present the highest rate of attempted suicide among all psychiatric illnesses: 20% of gambling addicts and 12% of gambling addicts' spouses attempt suicide, and 9% of gambling addicts with heroin addiction attempt suicide.
Mr. Reed also points to special attractions for children and classes to teach everyone, including youngsters, the basics of horse racing and the gambling that goes with it.
Society must balance benefits to be secured in revenue-raising programs with social costs that must be paid. For that reason, a wise and compassionate government should provide for treatment for those caught in the web of addiction.
ANOTHER RESOURCE . . .
Crain's story detailing Illinois insurance companies' use of credit history to determine potential risk in drivers merits amplification ("Insurers link credit woes, higher rates," May 15).
Property and casualty insurers have determined through exhaustive study that financial stability information is often a powerful indicator for potential loss. Insurance companies have always assessed customers to determine which drivers are better risks than others. The use of credit history has added to the companies' resources to assure that the best drivers will be accorded the lowest rates.
Illinois insurance companies using financial stability as a risk indicator do not assess creditworthiness to determine the chances of customers paying premiums on time. Rather, these criteria are now being used because studies show drivers with serious financial instability are 40% more likely to produce losses than drivers with good credit history.
I wish to clarify a detail in Crain's article regarding regulatory activity in response to some insurers' use of credit scores.
The article stated that the Illinois Department of Insurance took issue with plans written by Reliance and Travelers. The department has objected to certain credit-related filings, but the above-mentioned plans were accepted as originally submitted because they conformed to our requirement that, if an insurer uses credit as a rating tool, it must do so uniformly within a company.
Yes, North Michigan Avenue traffic is a nightmare.
Although a Lakeview resident and a transit user, I abhor having to go to or through North Michigan Avenue ("City's boom gets snarled in traffic jam," May 22).
The biggest problem? Pedestrians who ignore the "Don't walk" traffic signals, especially on the east side of the street. As a result, only a couple of vehicles get to make a right turn per light change while through traffic backs up, including buses that are supposed to be part of the solution.
What is a good, low-tech, partial remedy? Policemen with enforcement power who know how to direct traffic and control pedestrians. Unfortunately, some officers and most traffic aides do not.
| 2019-04-23T22:31:17 |
https://www.chicagobusiness.com/article/20000603/ISSUE07/100014386/letters-to-the-editor
|
0.979927 |
Maybe you had to get the mailing they sent with it to use it then.
tried the code and got the following message: Your account is not eligible for Promotional Code 'wkgr1'.
Wrong post... this is for jacamo!
B&Q voucher not available at present - "We are not able to respond right now. Please try this operation again at a later time."
You sure? It's the first thing that pops up when you navigate to their site. Are you a new customer?
just tried using and code is expired. tried to get another one from their facebook page it gave same code - looks like it has expired and they havent updated their facebook page yet.
Thanks :) Have added the start and expiry dates as given by Jacamo when I added the code.
crap company! don't use them! returned an item & never got a refund they added it to my jacamo account then deleted my account! lost £40 AVOID!!!!!
They've always refunded me when I've sent things back - never had any problems!
Worked for me today. Voted hot.
I recently got a xmas snow globe with one of my orders from jacamo haha how random!
Youo might be interested in the fact that certain shirts when ordered together net a £5.00 off, going from £12.50 to £10.00. Product codes IB512KS, IB513KS, IB514KS.
Yeah, I was eyeing up the Caterpillar jacket myself, and came over looking to see if there were any other codes.
Added dates per Jacamo site.
15% off Discount Voucher Code for 15% Off all first orders @ Jacamo!
Try looking for bargains rather than items we KNOW are always overpriced in catalogues, sheesh.
I have no idea about games, as I wouldn't ever have considered looking on a clothing site for stuff like that. My son is delighted with his three pairs of Lee Cooper Jeans, and two other pairs of jeans he got though, for £80 rather than £100. Hardly a rip off.
This deal would have expired by now stb17.
Have you bought from them before ramraideruk?
cex unlocked a, b or c mean?
Second sky q mini box - can i buy from ebay?
New contract but staying with same network and keeping my number?
Gta5 do i invest in anything before i start the Merryweather Heist?
2x 1st class stamps instead of large letter stamp?
Accidentally put a bit of unleaded in my diesel car, does it matter?
Handgun Shooting ranges in UK?
Driving Question: 4 people in the back only 3 seat belts?
are allbeauty.com reliable and are products 100% genuine?
What channels do you get on SKY with no subscription ?
Track you partners mobile phone for free!!!
Can you get Sky Go legit on Amazon fire tv stick?
Visit their online shop at www.jacamo.co.uk.
“Add to bag” the items you wish to order.
Select “Bag” or “Continue to checkout” to view the items in your basket.
In the box labelled “Add Promotional Code”, enter the voucher code you wish to use and click “Apply”.
The amount of the voucher will be deducted from your order total.
Click “Go to checkout” and sign into your account or register as a new customer.
Choose your delivery and payment details in the following steps and finish your order at jacamo.co.uk.
| 2019-04-20T20:43:23 |
https://www.hotukdeals.com/vouchers/jacamo.co.uk?page=4
|
0.999017 |
Beethoven: Piano Trio No.11 in G, Op.121a, 10 Variations On "Ich bin der Schneider Kakadu"
#11, opus 121a; Trio opus 38(after the Septet, opus 20); Trio in D (after Symphony #2; Triosatz in Eb; andTrio #4 in Bb opus 11 'Gassenhauer.
in the earlier trios the piano part was so written. Still one might compare this set with those of Kogan/ Gilels/ Rostropovich and Stern/Istomin/Rose.
I find Pressler the chief attraction; he plays with a beautiful tone, vivacity in the outer movements and sufficient gravitas in the slow movement.
behind them- it was an overwhelming experience. Even so, I would still award the Beaux Arts set five stars.
It is good to have this box, not only its great performances, but because it includes the WoO numbers also. WoO38 is a find.
Strongly recommended for serious listener and Beethoven disciples.
Any chance that this work can be recorded again with the same players but on original instruments?
In their 50th year now, The Beaux Arts Trio has achieved legendary status for their crisp, elegant, and polished playing and adherence to the spirit of the music. Pianist Menahem Pressler (the last remaining original member) finesefully drives the music forward with his deft and articulate keyboard work that is always marvelously alive and bouyant. His style has an attractive lyrical quality to it which blends marvously with the two strings whose playing is of the highest caliber. It doesn't seem the Beaux Arts Trio recorded many "duds" as far as I can tell, and this excellent 5-CD set is no exception. And while their classy playing might project a slightly more refined than dramatic Beethoven here, such a style is well suited to the classical-era style of these earlier works prior to 1805 and Beethoven's slow-but-sure departure from classical forms and mannerisms.
Part of this more 'classical Beethoven' found in most of the trios here is partly because of the genre. The piano trio genre was not the place Beethoven experimented or "went all out" - as he did in his symphonies, quartets and piano sonatas - but more a genre he wrote to sell music (although the empassioned C-minor trio of Opus 1 seems more written for Beethoven's own expression). In the late 1700's, the piano trio was a popular combination for the home music market for many a Vieneese amatuer. Also, it was common and in demand for composers and publishers to transpose popular orchestral works down to the piano trio (such as the two in this set) for this market of paying customers. What was also a trend then was including the themes of the latest "pop music" people were humming around town in theme & variation movements, which is seen here in some of the trios in this 5-CD set.
The quality of music Beethoven wrote in the trios is very good and highly enjoyable to listen to. Surprizingly, some of Beethoven's finest piano trios are found in his Opus 1 set - the very first compositions Beethoven found worthy of publication. These were written in his early 20's and already show the musical skills and depth of a fully-mature composer. Highlights of these recordings are the attractive and musically-rich Opus 1 set mentioned, the magnificant opus 70 two trios, and the noble, heroic-period "Archduke Trio" - perhaps the most famous of them all. The Beaux Arts Trio not only brings out the power and drama of the fast movements quite well (a stunning C-minor Op.1 no. 3), but especially conveys a most attractive lyrical quality in the slow movements with great beauty and sensitivity, such as in the haunghtingly attractive Largo of the famous "Ghost" trio, (Op. 70 no. 1).
This is a really high quality and distinguished set of recordings packaged nicely in a super-slim box for easy storage and offered at an attractive price. Bottom line is they play this music superbly. Another Beaux Arts Trio set that is equally staturesque and warmly recommended is their 9-CD set of Haydn's Piano Trios (Philips). Compositions - 5 stars; Performance - 5 stars; Sound quality - 4 stars.
When I changed my computer's audio player from Windows and/or iTunes to Winamp, I realised that my MP3 download of the "Archduke" trio wasn't good enough.
So I went browsing through Amazon's audition tracks. You'd be amazed how many interpretations there are of the "Archduke". You'd also be amazed at how poorly some of the players are served by their recording engineers. Instruments are allowed to dominate in a manner that simply should not happen in chamber music. It is, after all, supposed to be a conversation between intelligent people, but some recordings are virtually piano solos.
In the end, that was why I bought this set. The players work together extremely well, and their recording engineer works with them extremely well, so the result is chamber music as it should be.
Unfortunately the recordings are about 30 years old, and it shows. In particular, the piano can be quite harsh. I don't like some of the tempi either, but that's a matter of personal preference. I think they're a bit slow. In the case of the "Archduke" first movement, very slow.
| 2019-04-20T10:19:10 |
https://www.amazon.co.jp/Piano-Trios-Beaux-Arts-Trio/dp/B00005ND43
|
0.999352 |
Think you know everything there is to know about the world's tallest peak? Think again! We have seven little known facts about Mt. Everest that are sure to give you a new perspective on this iconic mountain, which remains an alluring destination for adventure travelers, trekkers, and climbers even in the 21st century.
Just How Tall is Everest?
Back in 1955 a team of Indian surveyors visited Everest to take an official measurement of the mountain's height. Using the best equipment of the day, they determined that it stood 29,029 feet (8848 meters) above sea level, which remains the official altitude recognized by both the Nepali and Chinese governments to this day.
But, in 1999 a National Geographic Team placed a GPS device on the summit and recorded the altitude as 29,035 feet (8849 meters). Then, in 2005, a Chinese team used even more precise instruments to measure the mountain as it would stand without the ice and snow that have accumulated on the summit. Their official measurement of just the rock itself came in at 29,017 feet (8844 meters).
Which one of these measurement is correct? For now, the official height of Everest remains 29,029 feet, but plans are afoot to measure the mountain once again, particularly since it is believed that the height may have changed following the 2015 earthquake. Perhaps we'll finally get a consensus of the true height at long last.
The first successful summit of Everest was recorded by Edmund Hillary and Tenzing Norgay on May 29, 1953. But, there are some who believe that it was actually climbed much earlier.
Back in 1924, an explorer by the name of George Mallory, along with his climbing partner Andrew Irvine, were part of an expedition attempting to complete the first ascent of the mountain. The duo were last seen on June 8 of that year just below the summit but making steady progress upwards. Shortly thereafter, they simply disappeared, leaving behind a mountaineering mystery for the ages. Did they actually make it to the top nearly three decades before Hillary and Norgay or did they perish somewhere below the summit?
In 1999, a team of climbers discovered Mallory's remains high on the slopes of Everest. The body did little to reveal whether or not he actually reached the top and unfortunately the team's camera was not found amongst his gear. It is believed that Irvine was actually carrying the camera when they made their ascent, and that device could hold the photographic evidence of their success or failure. To date, Irvine's body – and the camera – has not been found, but if it is ever uncovered, it could potentially change mountaineering history forever.
Who Has Climbed Everest the Most?
Climbing Everest is no small feat, and reaching the top remains a tremendous accomplishment. But for some people, climbing the mountain once just isn't enough. In fact, a climber by the name of Kami Rita Sherpa has been to the summit on 22 separate occasions, giving him the record for most successful attempts on the mountain. Mountain guide Lhakpa Sherpa holds the record for most summits by a woman, having climbed to the highest point on the planet nine times.
The record for most summits by a non-Sherpa is held by American Dave Hahn, a guide for RMI Expeditions. He's made the trip to the summit 15 times as well, which is an impressive number too.
For most climbers, reaching the summit takes several days with stops at various campsites to rest and recover along the way. But a few talented alpinists have been able to go from Base Camp to the summit in blazingly fast times, setting speed records in the process.
For instance, the fastest time for an Everest summit from the South Side in Nepal is currently held by Lakpa Gelu Sherpa who in 2003 managed to go from BC to the top in just 10 hours and 56 minutes. Lakpa spent a few minutes on the summit enjoying his accomplishment before turning back, completing the round-trip journey in just 18 hours, 20 minutes.
Meanwhile, on the North Side in Tibet, the record stands at 16 hours and 45 minutes and was set by Italian mountaineer Hans Kammerlander back in 1996.
In the Buddhist culture of the Himalaya Everest is known as Chomolungma, which translates to "Goddess Mother of Mountains." As such, the peak is seen at a sacred place, requiring all mountaineers to ask for permission and safe passage before they actually step foot on the mountain. This takes place during a puja ceremony, which is traditionally held in Base Camp prior to the start of the climb.
The puja is performed by a Buddhist Lama and two or more monks, who build an alter out of stones at the campsite. During the ceremony they ask for good fortune and protection as the climbers prepare for their ascent. They also bless the team's climbing equipment, including ice axes, crampons, harnesses, and so on.
For the Sherpa people this is an important step that must be completed prior to starting the expedition. Most won't even begin and Everest expedition without undergoing a puja ceremony first. Is this just a superstition? Quite possibly. But it is also a tradition that dates back hundreds of years and one that most foreign climbers are honored to take part in.
Age is just a number when it comes to climbing Everest. Sure, most of those who travel to the mountain are experienced climbers in their 30's and 40's, but others certainly fall outside that age group. For instance, the record for the oldest climber to ever reach the summit is currently held by Yuichiro Miura of Japan, who was 80 years, 224 days old when he topped out back in 2013. The youngest person to ever summit the mountain is American Jordan Romero, who accomplished that same feat at the tender age of just 13 years, 10 months, and 10 days in 2010.
Recently, the governments of Nepal and China have agreed to put age restrictions on climbers, requiring them to be at least 16 years old before attempting the mountain. Both countries have done away with a cap on the age, although more senior climbers may be required to pass a medical exam before starting their expeditions.
Sadly, Miura passed away on Everest in 2017 while attempting to reach the summit once again at the age of 85.
While the summit of Everest may be the highest point on the surface of the Earth, it isn't actually the tallest mountain on the planet. That distinction goes to Mauna Kea in Hawaii, which is actually 33,465 feet (10,200 meters) in height, a full 4436 feet (1352 meters) taller than Everest.
So why isn't Mauna Kea recognized at the highest peak instead? Because most of the mountain actually sits below the surface of the ocean. Its summit only rises 13,796 feet above sea level, making it appear to be relatively modest in size when compared to the Himalayan giants.
These 23 Facts About Brunei Will Bewilder You!
| 2019-04-21T04:49:11 |
https://www.tripsavvy.com/little-known-facts-about-mt-everest-4136176
|
0.999321 |
How did God know the only way we would feel totally transformed is if we started from scratch again? When the Bible talks about how we are a new creation in Christ Jesus He is not talking about it in a literal sense. More specifically we are not re-born to our earthly mothers but reborn spiritually. However, I think it's an interesting concept. When we want to start a new adventure, a new relationship, or just a new way of life we use this term of starting over.
What does that look like to you?? Some people see it as a chance to get rid of all their earthly possessions and just buy new that coincides with their new life. Others may not view this earthly possessions more than just things they need so they do not feel as tied to them emotionally. What kind of person are you in this scenario?
I find myself in the middle. I am not tied to any particular item. Yes they would be nice to have so I won't have to go out later and rebuy it. However, some people feel a deep sentimental value to that item and want to get rid of that item because they think it will get rid of the memory.
Whether that person destroys their own personal property because they think it will help them heal or cleanse their life of a particular person--I find it odd myself they think by getting rid of a physical decor that they will feel better and have closure over that relationship.
| 2019-04-22T06:52:20 |
http://theheartofachristian.blogspot.com/2012/04/new-creation.html
|
0.999836 |
As the camera lenses get more specialised and differentiated for their characteristics it seems good idea to look at the test results in context of methodology that was used for obtaining the data. Here I sum up a few more or less known points about the topic.
The point of focus method tests for optimal resolution at points of focus. In order to record it, it refocuses the lens at different positions in the image field and measures resolution only at the points of focus. It shows maximum resolution for a given spot of the field the lens is capable of at the position. These values are then plotted into resolution map of a lens and can be compared to results of other lenses. Observations made during the test about impacts of aberrations such as more pronounced field curvatures are usually commented on separately by the tester.
With most other tests the point of focus is in the centre and these would be the standard measurement which maps aberrations as they develop with growing distance from the centre and combines it with a lens resolution. It gives us idea of resolution of the whole image and thus separates sharp plane of focus (and aberrations free areas) from the areas that are less sharp or downright soft. Comments on how the field curvatures observed during the testing impact the resolution are often made as the curvature isn't directly apparent.
Plots for field curvatures (Petzval and astigmatism curves) have to be measured and presented separately. That is common with the later but it is rare to see the charts for the former (AKA curvature of field).
Both methods show something different and it is not difficult to see why the ‘standard one’ would be considered a complete representation of the characteristic of the optic. But, and this is not surprising, the method has its limits or imperfections and those are addressed by the more specialised and resolution-biased ‘refocusing method’ which can often be more relevant for the real world photography, or, more precisely, for many kinds of it. Particularly so when you want to know what resolution at the point of focus you are likely to get which will also be related to the way the photographer is focusing, such as moving the camera focus point so that the focus is precisely on the subject at the time of the exposure with no concerns over what resolution in other parts of the frame, e.g. in corners, is achieved. The focusing methods also develop with camera technology which makes them more efficient and convenient to use. On the other hand, when we need to consider even plane of sharp focus we are looking to see how the lens does in this respect and standard measurement aims to cover just this. It gives us a more complete picture about the lens. Despite the fact that its limitations mean that establishing the precise point of focus resolution away from the centre is not possible.
Nikon has recently signaled prioritising lenses optimised for point of focus applications in connection with the common use of video and emphasis on their own way of expressive rendering. This is a welcome step for differentiation in lenses and their use today. Using today’s means for this design, it also deals with various aberrations (along with the other ways) by making them work towards the design goal which can even be cost and weight effective. As this approach represents more subjective renditions in the lens' output it is vital to get the cocktail of the aberrations balanced well and in line with the raising standards for quality of the photographic lenses. The differentiation means that lenses that are corrected (to some degree) for even plane resolution are also included for the variability of use and prioritised for suitable specifications and applications. It would otherwise mean limiting production to only a certain part of the market while including both design philosophies promises expanding production possibilities and serving even wider needs of Nikon users. And this is what we can see happening. As the lens designs are taking advantage of the technological advances, Nikon has also raised the bar for the range of the optimisations involved and in how specific (effective) they are in shaping the individual lens so that it channels the advantage into creating distinct and unique Nikkor lenses in the digital era. This transfers, along with commonly known progress in the sensor and digital imaging, into new meaningful impulse to photography, its use, potential and as the photography changes just as we do, its meanings.
| 2019-04-18T16:38:02 |
https://www.sunwaysite.com/Zlatka-Photography/Notes-Point-of-Focus-Even-Plan/n-BKx6Sj
|
0.99936 |
Did I make the right decision? Should I have thought longer and harder and made a better one? It doesn't matter if it's making a smaller choice or resolving a big problem, I keep second-guessing myself... worried, feeling small, insignificant, and insecure.
I struggle to figure out if I'm good enough. Struggling with doing the right thing, always second guessing myself, filled with self-doubt.
Perhaps, it has a lot to do with running a service business. I am always required to be on my toes, having to please everyone, having to always be at my best. I'm not worried so much about the competition as I am worried about doing something stupid or inadequate.
Dealing with issues of self-confidence is a common problem among entrepreneurs I've dealt with over the years, especially those of us in the service sector. If left unchecked, this nagging self-doubt can keep gnawing away at you.
I believe that insecurity can be a positive. You can embrace it and use it as your driving force.
You don't have to spend a second trying to overcome it, if you simply wrap your arms around it and make it your best friend. When you leverage your insecurity and use it to your advantage, it can actually help you excel.
Sometimes a little self-deprecating humor can break the ice. Making a joke about yourself helps you and others laugh about any insecurity, before it becomes a negative. Try "Ugh! I always do that," instead of freaking out over a small mistake or oversight. You'll be surprised how everyone (especially you) becomes less tense and up-tight.
1. Fill your mind with positive thoughts. It helps to stay away from grumpy colleagues and clients, and turn off negative news reports.
2. Let go of small mistakes. Try not to dwell on small errors. Don't sweat the small stuff. Instead, take responsibility, apologize, fix the mistake as soon as possible and move forward.
3. Continue to grow and improve. A small accomplishment can help boost your confidence, even if it's not entirely related to your business. Learn a new skill, take a class at a local university or read a book on a subject that interests you.
4. Take baby steps. It's easy to face self-doubt when you try little steps before you commit to major ones.
5. Relax. Take some time off and schedule time to play. Get away from your challenges.
6. Ask for advice. Call a trusted friend, advisor or colleague and get his or her best advice. Often an objective opinion will help you look at life differently, overcome your challenges, and transform your attitude.
7. Stop seeking reassurance. Commit to your decisions and go with them.
8. Stop trying to be too perfect. Try adjusting your standards from perfect to good enough. That is usually all that people expect from you. Admit it...you can never know everything, but neither can anyone else.
Whenever you have trouble making a decision, remember that few decisions are irreversible or catastrophic. Even a tattoo can be removed, right?
So there are no wrong decisions, because each one gives you an opportunity to learn and adjust your path in the future. A bad decision can rattle your self-confidence, but you will soon be able to get back on track.
Over time your will realize, as I did, that indecisiveness is worse than being wrong. It's a difficult lesson to learn...one that took me a long time to realize (and one that I sometimes forget). Any decision-maker needs to lead and not be wishy-washy.
When we find and speak our voices, we feel better connected to ourselves and others, and we have an important tool for navigating daily tasks and decisions, big and small. And our self-confidence soars!
| 2019-04-22T11:23:47 |
https://www.huffpost.com/entry/struggling-with-selfconfi_b_8089172
|
0.999999 |
Is it possible to set a global rule which will block all traffic from the computer other than that which goes through the VPN connection? There is a setting for global rules in comodo, but attempting to do it in the same way as for utorrent doesn't seem to work, as it blocks the VPN itself. Basically I'd like to have it set up so that the computer can't connect to the internet at all except through the vpn.
Yes, it is possible. Remember to allow packets for your physical network card from and to the entry-IP address of the VPN server you're connected to, otherwise you will block every and each packet as you have experienced. There are several ways to accomplish this. An example is to block everything from and to your network card (NOT going to and coming from the entry-IP address) AND (NOT coming from or going to your TAP-Win32 adapter).
That's great, thanks! How might I go about doing that in comodo (or otherwise)? I've looked through the various global and other ruleset settings, and there's no obvious option to make rules involving my network card or TAP-Win32 adapter. Apologies if I've just overlooked it.
With Comodo, first detect the Network Zones corresponding to your adapters (usually Comodo gives them names like Home #1 etc. according to your preferences). Then apply the global rules to those network zones (both when they are the target and the sender) in the tab "Global Rules".
You'll need to know the Network Zone corresponding to your local adapter (10.4.0.0/8) and to your physical adapter (WiFi or Ethernet card). See the tab "Network Zones" inside the "Network Security Policy" menu of the Firewall section in order to detect them.
The Network Zone corresponding to the TAP-Win32 adapter can be authorized to receive and send any packet from any source or to any destination, from/to any port from/to any port.
The Network Zone corresponding to your physical adapter must allow ONLY the packets coming from or going to the entry-IP address of the VPN server or the TAP-Win32 adapter.
If I'm reading your reply correctly, then all I have to do is go to global rules and set the network zone corresponding to my physical adapter (wifi card in this case) to allow only traffic from the VPN entry IP address or the the network zone of the TAP-Win32 adapter.
Comodo firewall currently has five network zones listed: one called "Loopback Zone" and four called "Home #1"," Home #2", "Home #3", "Home#4". How do I tell which corresponds to my physical adapter and to the TAP-Win32 adapter?
Ok great, I have the IP addresses of my physical adapter and TAP-Win32 adapter.
Also, will the VPN entry IP address be the same each time I connect, assuming that I always use the same airVPN server?
- (only if already connected): open Comodo Firewall "View Active Connections" and see which IP address openvpn.exe is exchanging packets. That IP address is the entry-IP address.
OK, I've now got the various network zones figured out. In this case the Physical adapter (wifi card) is Home #1 and the TAP-Win32 adapter is Home #3. The VPN entry address appears to be the same as that of my wifi card.
I've tried creating various rules in the "Global rules" tab to match what you've written in your post, but I can't seem to work it out.
What rules do I need to create?
I've tried everything said here, as well as the other thread. But the only solution that works for me is still blocking individual applications.
Any other workable solution to achieve no possible traffic outside of the VPN?
Yes! a guide would be awesome, I'm liking this vpn a lot, a bit tricky at first ( I'm an idiot). But I also would prefer to stop all incoming and outgoing traffic unless I'm on the VPN. If it's possible this thread should be stuck at the top "Sticky" and solved?
I messed around with Comodo Firewall's settings for a bit and I think I found a solution.
Comodo Firewall -> Firewall -> Network Security Policy -> Global Rules -> Add.
For connecting to AirVPN with the AirVPN client, I've tried excluding AirVPN.org's IP address in the rule, but I was still getting the same error. Is there a way to log into the client without having to disable the global rule? Also, is there a difference between blocking IP and blocking TCP&UDP/IMCP?
In this way airvpn.org will be resolved without the need of a DNS query outside the tunnel which is correctly blocked with your rules when you are not connected to an Air server. You will still have to authorize packets from and to 46.105.19.36 in the firewall.
Of course if we change the IP address of our frontend you will have to update your hosts file.
Hello. I have been working on this for awhile and Air hasn't really put out a tutorial on it despite so many clients being interested. Thank you greg for the post! Here is what I have in Comodo.
Why can't you exclude the range 10.4.0.0 - 10.9.255.255 instead of the entry address of an individual server? Is there a way to exclude all servers IP's at once? When I try to allow the range it logs into the Air client but doesn't connect to any servers. With the single entry IP I can only connect to the one server I have listed. So close to working!
To be honest, I had no idea what I was doing and was just messing with the settings while following what Admin said in page 1.
I have been trying to add separate rules and it won't allow me to connect after adding rules for more than 1 server. It's confusing me, it doesn't make sense.
Did you add the new rules on top of the block rule?
I added an "Allow IP In/Out From MAC Any To IP <Entry address> Where Protocol Is Any" rule for each server and I have no problems connecting to any of them.
Thanks for the screenshot. I think I got it working, but how did you allow "all"? Probably doesn't matter but just curious. Also what is [AirVPN - Sirius], just 108.59.8.147 I'm assuming. Here are all the IP's you need in text format for anyone else working on this. Cheers.
That's one of the default rules that Comodo sets to your network zones. The description says all but if you set the same rule without a description, it will say "Allow IP In/Out From MAC Any To In [AirVPN - Sirius] Where Protocol Is Any." I wanted to keep at least one of the default rules in case if I screwed up by deleting those rules, so I kept that rule and renamed the description to [AirVPN - Sirius] to match the network zone's name. Probably should have deleted that rule before taking that screenshot, but I guess it's a bit too late now.
Ah okay I see. Another question, sorry I am full of them. Now I want to stream TVersity to my XBOX, but obviously it won't because I have everything blocked when that computer isn't connected to the VPN. When connected it does the same thing. What global rule would I add to allow this to stream? I tried allowing 192.168.1.1-192.168.1.255 but it didn't work. Both IP and UDP/TCP. Thanks dude you are awesome.
any update on the list or anything else anybody would like to add?
No it's not, that's the exit IP.
| 2019-04-20T22:54:03 |
https://airvpn.org/forums/topic/2183-blocking-all-non-vpn-traffic-windows/
|
0.854336 |
Dems don't hate corruption, they want in on it.
We don't have to wait any longer. We now know that Nancy Pelosi's Democrats are a bunch of crooked politicians, just like the ones they replaced last November.
Ms. Pelosi became Speaker of the House partly because voters were upset at Republicans stealing tax money. She vowed to change the laws to "clean up Congress." What happened? Some of the names changed, and lobbyists can't take Congress-critters to dinner any more, but legislators can still earmark money for favored companies. They even voted down a law that would have made it illegal to earmark money for companies where their relatives work!
The only thing Democrats didn't like about Republican corruption was that Democrats weren't in on it. Now that they have the power to steal our money, the sky's the limit, assuming there's any limit at all.
The men who wrote our constitution knew that only government has the power to keep government under control so they set up a system of checks and balances. Congress is part of the legislative branch of government, they're supposed to blow the whistle when the executive branch gets out of line. Similarly, the "justice" department works for the executive branch, it's supposed to put crooked legislators in jail.
This worked pretty well for 200 years, but our government now takes in so much money that both branches got greedy. There seems to be a "don't ask, don't tell" deal between Congress and the administration where they don't rat each other out for stealing our money.
Look at Iraq. There's been talk of billions of dollars in reconstruction funds going missing. Has Congress taken action? No, if Congress complained how the administration wasted money, the administration might criticize them for the way they waste money.
For example, the administration is wasting hundreds of millions of dollars on a system to computerize military medical records. They've been at it for years with basically no progress. The head of the project visited the Washington Medical Center where a private company had spent less than $20 million building a system which would do the job, but he wasn't interested. All this money down the drain and where's Congress? Counting earmarks.
Congress and the administration have decided to let each other steal.
Where are the newspapers which Thomas Jefferson charged with keeping both parts of the government in line? The mainstream media hate President Bush, why don't they go after the money wasted in Iraq? They act like they're afraid they might damage their Democrat friends if they got the taxpayers looking too closely at how tax money is thrown away.
Maybe taxpayers are smarter than politicians think. Mr. Bush has very low approval ratings, but Congress' ratings are a lot lower. Hopefully, they'll vote some of the rascals out this time.
| 2019-04-25T15:51:50 |
http://www.scragged.com/articles/nation-unbuilding-the-decline-of-checks-and-balances
|
0.991166 |
How many people / food could I cook in a 5 litre capacity water bath?
Sous vide rooky here looking to explore the field a bit. I have an opportunity to get a good professional 5l water bath for not too much money, around half the price of a Sousvide Supreme. I was wondering how much food I would be able to cook with 5 litres. Would I be able to do dinner parties for say 4 or 5 people or should I really think about something larger?
Your expertise is much appreciated!
5 liters of water is not very much. Why don't you put 5 liters of water in a pan and see what it looks like? Then imagine whatever food you are going to cook being put in that amount of water. Also, will your apparatus also take up room in your cooking vessel? I have been cooking sous vide since last Christmas and I prefer to have enough water in my pot so that the water can circulate easily amongst the food pouches.
That sure looks like an interesting water bath. Does it have a lid? I didn't see a price listed - do you mind me asking how much you paid for it? I love my sous vide set up - I have cooked quite a few things in it so far. I especially like steaks and scallops done sous vide. You can also make a killer Hollandaise sauce sous vide. In fact, it is foolproof. Have fun with it and please post a few of your experiments!
Doesn't it? Very mad scientist. It cost £135 from ebay: http://cgi.ebay.co.uk/ws/eBayISAPI.dll?ViewItem&item=271057308967&ssPageName=ADME:L:OC:GB:3160. Has a loose fitting lid apparently but I'm sure I should be able to fix that. Thanks! I'm really looking forward to using it. Do you recommend a sealer as a necessity or can I get away with ziplocks and squeezing the air out?
It doesn't matter if the lid is loose, as long as it fits. It is only there to keep the water from evaporating. I have a Polyscience Immersion Circulator that I use with a couple of different sized stock pots, depending on what I am making. I use plastic wrap across the top of the pot to keep the water from evapourating.
I use a Foodsaver to vacuum pack my pouches. I have a setting on it that lets me seal stuff that I don't want to have squished, such as fish. However, zip locks work if you use the water displacement method. One thing I learned about that method is to place the bag you are going to cook the food in, into a bigger bag. That makes using this method easier.
Along with this website, you might also try eGullet. There is a wealth of info there on sous vide. Sous Vide Supreme and SV Cooking is another one. If you google Douglas Baldwin, he wrote a book on sous vide and has some demo videos. Also, Jason of this site has a couple of good books and in addition, will often answer questions. And above all, don't be afraid to ask questions. I sure asked a lots of them and folks are only too glad to help.
I take it you are in England?
Sorry for the late reply. Wow, a Polyscience eh. Very nice. Love the look of those but not ready for that kind of outlay, or indeed how much I'm actually going to use it in practice. While I love food and cooking I am naturally remarkably lazy and rather disorganized... Thanks for the tips! I think I'll start with zip locks and see how it goes. Presumably when you say double bagging, the second bag has to be sealed with water displacement as well. Good sites as well. Recently disovered eGullet, looks great! I am indeed. London based actually.
| 2019-04-25T00:15:35 |
https://www.amazingfoodmadeeasy.com/modernist-cooking-forums/sous-vide-equipment-forum/topics/how-many-people-food-could-i-cook-in-a-5-litre-capacity-water-bath
|
0.999986 |
Olivia Newton-John used ''medicinal cannabis'' to ease the pain of her breast cancer treatment.
The 68-year-old singer is currently battling cancer for the second time, and has now said she finds the drug - which is legal in the state of California - ''really important'' when it comes to easing the pain which comes alongside the aggressive cancer treatment, as well as promoting ''healing''.
Meanwhile, Olivia recently revealed she is ''feeling great'' since announcing her second cancer diagnosis earlier this year.
Olivia - who previously battled breast cancer in 1992 - was told her cancer had metastasised after suffering severe back pain earlier this year from what she thought was sciatica.
Doctors then informed her that her cancer had spread to the sacrum, which is the triangular bone in the lower back, situated between the two hip bones of the pelvis.
In a statement released in June, Olivia said she was ''totally confident'' that she would beat her cancer a second time, and hoped to ''inspire others'' with her success story.
| 2019-04-19T20:29:20 |
http://www.contactmusic.net/amp/5805488
|
0.999744 |
Q:? How can I change Product key in Visa/ 7 Operating systems.
To Change a Product Key in Vista/ 7 OS Follow these simple steps.
4. In the bottom where it says Product ID next to that you will find a link which says Change product key, click on that.
2. right click on My Computer and select Properties.
3. In the bottom where it says Product ID next to that you will find a link which says Change product key, click on that.
3. In the System Properties window, on Bottom-right corner, there is option to change the Product Key!
Windows XP Professional OEM on Windows XP Professional Retail.
Since Windows XP is no more supported it is best advised to upgrade to Windows 7 or Windows 10 Operating system for better protection and regular security updaets.
| 2019-04-20T10:21:36 |
https://www.winkeyfinder.com/faqs/how-can-i-change-product-key-in-visa-7-operating-systems.php
|
0.999508 |
While this works nicely, it has the disadvantage of requiring another wire. Can we do it without introducing any additional connections?
I accidentally stumbled upon another way: When the WS2812 LEDS are set to the brightest level of white (255,255,255) they draw around 60 mA of current. This leads to a measurable voltage drop on the power rails. So we can get a good idea of the number of LEDs connected to a power source by setting it to a known pattern and measuring the resulting voltage drop.
Many microcontrollers actually allow using their internal ADC to measure the supply voltage. If the MCU is connected to the same power supply as the LED string, it can determine the current draw of the WS2812 LEDs without any additional hardware.
I tried this on an ATtiny841. The code below uses the internal ADC to measure the band gap voltage (1.1 V) using VCC as reference. From this, it is possible to calculate the actual supply voltage. It is necessary to average a number of measurements due to noise introduced into the power rail by the WS2812 pulse width modulation.
Measures the supply voltage internally using the ADC.
The image below shows the output for a WS2812 string after converting the ADC output to a voltage. Both the LED string and the microcontroller were connected to a USB port as power supply.
It is easily noticeable that there are only 16 LEDs in the string as no further drop of the supply voltage occurs after this point. Interestingly, this information can also be used to calculate the effective internal impedance of the power supply. In this example it is 0.5 Ohm, assuming that every LED draws 60 mA. The main contributor to the series resistance is probably a poly fuse that is used to protect the USB port power rails.
The image above compares the behavior of several USB power sources. The USB battery has the lowest internal resistance, as indicated by the flat slope. In this case it becomes more difficult to resolve the contribution of individual LEDs and it is only possible to get an estimate of the string lengths.
In conclusion, even though it is a bit hacky, measuring the supply voltage allows to get a good estimate of the length of a WS2812 string without any additional connections.
Nice write up. This effort is an excellent example of thinking about using additional information that you can gather about your circuit. Perhaps you can apply a little more math and make the test slower but more robust with the battery.
I suggest toggling the first LED 5 (tweak to optimize) times and getting an average voltage delta for the on vs off state. Use a fraction of this delta as your test to determine if there if the next LED is present (we only want to know is there an actual delta that correlates with the next LED’s on state). The turn the first LED off and toggle the 2nd LED 5 times. If there was a 2nd LED, uses a fraction of that voltage delta as your test value for your next LED. Obviously there should be no delta for an address with no LED. In essence we want to change from an analog mapping measurement into a digital comparison test with a shifting window level for the comparison.
Please let me know if you get a chance to try it. Great blog.
thank your for your comment. Yes you are right, there are several ways to make this more efficient.
By integrating more measurements it would be possible to get a higher resolution.
I agree that just using a single LED to “probe” could be more efficient, especially if there as a danger of overloading the power source. One could use the first LED to get the voltage delta, as you described, and then use a binary search pattern to speed up the process.
Indentation is off in the second code snippet.
of course, this is a little bit slower than your method.
This (or one of the single-led-at-a-time variants) should be fairly reliable for short strings powered by a relatively high effective resistance power supply.
It may be harder for a series of 150 or 300 LEDs. First off, they are going to require a much lower impedance power supply, and secondly the delta current (one lit LED plus N dark currents, vs no lit LEDs plus N dark currents) is going to get less dramatic (lower S/N). (and its worse still if you light the first K pixels rather than just one).
One could partially counter the latter effect by lighting M LEDs at a time rather than just 1, albeit one would get a less precise count that way (eg: lighting pixels 95-100 drops the voltage vs non lit, but lighting pixels 101-105 does not – so there are around 100 pixels).
Still, it’s clever and and quite useful in some circumstances, thanks for the tip.
Another thought – if one wanted, one could have a series resistor between the power supply and the load, which could be switched in or out of the circuit – eg: using a relay to short across it when operating normally (ie: many LEDs lit at once which would drop too much across the resistor)..
Rather than measuring the uC’s own power, one could use a resistive voltage divider (after the series resistor in the power line) to bring the measured voltage down to less than 1.1v and do the ADC against the 1.1v reference. This would allow finding the last pixel of even longer strings, and could work when the uC is using a separate supply or using onboard regulation from a higher Vin, and when the LEDs are running at 12v, etc – at the cost of 3 resistors and a relay.
Again, that’s a slightly different niche than your no-external-components approach. Like a control box into which different strings might be plugged. And if you wanted to be fancier than that, you could use a current sensor chip for more accuracy (and no need for a relay).
I am a couple of years late here… But… Can you help me decode the below technique you mentioned? Would I wire in the DO pin of the last LED into a digital input on the arduino? What am I reading here?
>An intuitive approach to counting the number of LEDs is to feed the output of the last LED of the string back into the microcontroller. According to the WS2812 protocol, the output of LED number n is low until data for n+1 LEDs has been pushed to the string. To count the number of LEDs, data is fed to the string until a rising transition is detected on the output. A relatively easy way to implement this is by using a pin change interrupt on the AVR, or a similar feature on other microcontrollers.
| 2019-04-20T03:39:44 |
https://cpldcpu.wordpress.com/2014/11/16/ws2812_length/
|
0.998732 |
To really understand the push-pull over the bungled talking points in the wake of the Benghazi attack, you have to understand the nature of the U.S. presence in that city.
Officially, the U.S. presence was a diplomatic compound under the State Department's purview.
"The diplomatic facility in Benghazi would be closed until further notice," then-State Department spokeswoman Victoria Nuland announced last October.
But in practice - and this is what so few people have focused on - the larger U.S. presence was in a secret outpost operated by the CIA.
About 30 people were evacuated from Benghazi the morning after the deadly attack last September 11; more than 20 of them were CIA employees.
Clearly the larger mission in Benghazi was covert.
The CIA had two objectives in Libya: countering the terrorist threat that emerged as extremists poured into the unstable country, and helping to secure the flood of weapons after the fall of Moammar Gadhafi that could have easily been funneled to terrorists.
The State Department was the public face of the weapons collection program.
"One of the reasons that we and other government agencies were present in Benghazi is exactly that. We had a concerted effort to try to track down and find and recover as many MANPADS [man-portable air defense systems], and other very dangerous weapons as possible," former Secretary of State Hillary Clinton testified before Congress in January.
The CIA's role during and after the attacks at the diplomatic post and the CIA annex in Benghazi have so far escaped much scrutiny.
The focus has been on the failure of the State Department to heed growing signs of the militant threat in the city and ensure adequate security, and on the political debate over why the White House seemed to downplay what was a terrorist attack in the weeks before the presidential election.
But the public needs to know more about the agency's role, said Republican congressman Frank Wolf, of Virginia.
"There are questions that must be asked of the CIA and this must be done in a public way," said Wolf.
Sources at the State Department say this context explains why there was so much debate over those talking points. Essentially, they say, the State Department felt it was being blamed for bungling what it saw as largely a CIA operation in Benghazi.
Current and former U.S. government officials tell CNN that then-CIA director David Petraeus and others in the CIA initially assessed the attack to have been related to protests against an anti-Muslim video produced in the United States.
They say Petraeus may have been reluctant to conclude it was a planned attack because that would have been acknowledging an intelligence failure.
Internally at the CIA, sources tell CNN there was a big debate after the attacks to acknowledge that the two former Navy SEALs killed – Tyrone Woods and Glen Doherty – were CIA employees. At a 2010 attack in Khost, Afghanistan, when seven CIA officers were killed in the line of duty, the agency stepped forward and acknowledged their service and sacrifice. But in this instance - for reasons many in the Obama administration did not fully understand - it took the CIA awhile to "roll back their covers." Petraeus did not attend their funerals.
Wolf said he and his office are getting calls from CIA officials who want to talk and want to share more.
"If you're 50 years old and have two kids in college, you're not going to give your career up by coming in, so you also need subpoena power," said the Republican congressman. "Let people come forward, subpoena them to give them the protection so they can't be fired."
But is the secrecy surrounding the CIA's presence in Benghazi the reason for the administration's fumble after fumble when trying to explain what happened the night of the attack?
There were 12 versions of talking points before a watered down product was agreed upon– suggesting an inter-government squabble over words that would ultimately lay the blame on one agency, or the other.
Perhaps the State Department did not want to get in the line of fire for a CIA operation that they in many ways were just the front for, the CIA "wearing their jacket," as one current government official put it.
The CIA did have an informal arrangement to help the mission if needed, but it was not the primary security for the mission. The State Department had hired local guards for protection.
People at the CIA annex did respond to calls for help the night of the attack. But despite being only a mile away, it took the team 20 to 30 minutes to get there. Gathering the appropriate arms and other resources was necessary.
None of this diminishes questions about how the White House, just weeks before the presidential election, seemed to downplay that this was a terrorist attack. Or the State Department's initial refusal to acknowledge that it had not provided adequate security for its own officials there.
But the role of the CIA, its clear intelligence failure before the attack, and - as it continued to push the theory of the anti-Muslim video - after the attack, bears more scrutiny as well.
| 2019-04-22T02:34:48 |
http://thelead.blogs.cnn.com/2013/05/15/analysis-cia-role-in-benghazi-underreported/
|
0.999647 |
In previous articles, I provided information encouraging specialty graphic printers, or print service providers (PSPs), to take initiative in determining what equipment to buy.
By educating themselves about the issues and the processes involved in design, print and finishing and reviewing their customers' needs, a PSP can make its own decision rather than rely on the recommendation of a single vendor. The goal of the article was to ensure that printers developed a plan for how to meet all of their customers' requirements up front.
I also suggested purchasing a less expensive flatbed printer paired with a digital finishing system rather than purchasing a printer with the maximum speed and quality that is affordable. I provided comments from some users who said that having both a printer and a finisher was economically the best decision they ever made. With this dual capability, virtually no finishing job had to be turned down. Additionally, it significantly increased company profitability as it broadened their customer base and did more for each customer.
In this article, I take the discussion a step further: A digital finisher is not enough in itself. While finishers are easy to use and do a great job with timely and accurate finishing, are your estimators capable of accurately determining the true costs associated with finishing? Based on my experience, this is still more of an art than a science and most printers do not estimate finishing very well.
Over the past few months, I have spoken with a number of estimating and workflow system manufacturers/sellers about how they estimate finishing. I learned that the quick answer is always: "We have an estimating component for finishing in our solution." But when pressed, the approach they take is simplistic. In fact, they often asked me: "How do you charge for finishing?"
One problem is that you may not get accurate results from your current estimates even though you rely on estimating software. Although you know how much your hourly costs are, you may not know how to accurately estimate the time and processes involved in finishing a group of complex graphic shapes nested on a substrate while minimizing both ink and material cost.
When I discuss finishing with PSPs, they admit that their estimate process may be off by a factor of three or more. This impacts both the profitability of finishing and the entire job. It negatively affects scheduling, causing problems with employee hours or even with meeting deadlines for other customers. If a job that was scheduled for two hours takes six or eight hours to complete, it can have a large impact. If you normally estimate $100 per hour for finishing, an extra four to six hours adds $400-600 to the costs -perhaps more than your expected profit. Also, the delay might require you to select a faster shipping method to meet the customer's delivery deadline and fast shipments may add $100-200 to the job.
And if you usually make more on finishing than printing, this hurts your overall profitability. Welding and grommets for banners, pockets and slits for folders and other similar finishing costs are addressed accurately by estimating programs. For digital finishing, however, estimating programs use customer-defined costs per square foot regardless of materials used, thickness/hardness of the material and complexity of the required cutting. What is missing is any kind of true analysis of the job that can make the estimate more accurate.
Recently, I spoke to a local wide-format printer to see if anything has changed. He told me that they always price finishing conservatively because they are often wrong. If the job is small, adding a "fudge factor" that increases the price may not jeopardize the job, and overall, may save their skin. When the job requires a large amount of finishing, they cannot add too much to the price, or they may lose the entire job. They need a good estimating tool for finishing and have not found one.
I am not sure why an accurate, reliable estimating system is not available. However, there are many more parameters to deal with in finishing than in printing. Print estimating is a well-understood process, and many available solutions can accomplish this effectively and reliably. Many print estimating products are being augmented to handle wide-format printing. RIP companies have added estimating to their repertoire, giving PSPs one tool to use from design through production. So, can the same level of precision and automation be accomplished for digital finishing estimates?
Unlike print, which is largely defined by known parameters such as image size, colors/varnish and ink density, it is difficult to "guess" the accurate cost of the finishing production work for many digital cutting/routing projects. And, it is particularly difficult when a customer asks for a project quote based on a rough pre-printing design concept. By gathering information from the rough design, putting it together with the knowledge base that can be built into a cutting estimator and using five key data points (bounding box size, contour length, degree of shape intricacy, material type and material thickness), you can produce a consistently accurate cost estimate. With additional data, you can increase the level of the estimate from precise to perfect.
Part of the problem in estimating digital finishing has been the lack of an organized group dedicated to it; manufacturers have not been able to put their heads together and tackle this issue. In cases where a manufacturer was interested in doing this, it was difficult to determine accurately through measuring instrumentation how the cutting device would perform on many different shapes and materials.
Fortunately, there is a solution coming from a digital finishing service bureau that had five different kinds of X-Y finishing tables in use. Their facility included knife cutting tables such as vibrating knives, drag knives, creasing wheels and routing, routing-only tables and laser cutters. Based on six years of use and this array of equipment, comparisons between the original budgeted estimates and actual production costs were made. This information was analyzed and a set of algorithms was developed and beta tested in North America and Europe.
The testing showed that the cutting work time for a print project finished on a digital cutting system can be estimated easily and accurately even if you know only a few important elements of the project. In most cases, you do not have the final print as a guide because the estimate is completed before you get the print job. While you cannot cut one piece and measure the results before finalizing the price, you have the information you need with the specifications and shape.
If you had the completed prints in Figure 1 to guide you, estimating would be easier. It would be even easier with the prints in Figure 2.
Instead, what you usually have might look like Figure 3, a hand drawn car, often on the back of an envelope, generally a very crude version of the final drawing. Even with such a simplistic drawing, we know how to gather the necessary information for finishing.
With a firm methodology in place, a developer can now work with members of the software and hardware community to embed accurate estimating into their own unique solutions. The estimating and/or workflow program or workflow software that you currently use can be updated to include this technology. And, manufacturers of flatbed cutting equipment can bundle the technology directly into their current offerings as an added value, just as wide-format printer manufacturers bundle print estimating capabilities.
While this new technology may be the first solution, it may not be the only one. For now, it shows that PSPs, given the correct tools, can turn finishing from a cost center into a profit center that significantly improves their level of service to customers. I propose beginning an industry dialog between those interested in bringing this technology forward and discussing other new, innovative solutions.
| 2019-04-26T00:19:10 |
http://mailhost.signletter.com/finishing/articles/2010-08-02-TurningFinishing-FromCostCenterToProfitCenter.php3
|
0.999987 |
व्यावहारिक क्षमताओं में वह सापेक्ष स्थाई परिवर्तन जो कि प्रबलित अभ्यास का परिणाम होता है उसे क्या कहते है?
आठ वर्ष के सुधीर की मानसिक आयु दस वर्ष है उसकी बुद्धिलब्धि कितनी है ?
अल्बर्ट बण्डूरा निम्न में किस सिद्धान्त से सम्बधित है?
राष्ट्रवादी कवि रामधारी सिंह ‘दिनकर’ को ‘उर्वशी’ के लिए ‘ज्ञानपीठ पुरस्कार’ किस वर्ष प्रदान किया गया?
59.चराचरम् (जगत् ) में कौन सा समास है ?
‘अभ्यागत’ शब्द में कौन सा उपसर्ग है?
What kind of adjective is used in the following sentence?
How did a Luz Long help Josse Owens?
(1) Luz Long advised Josse to draw a line a few inches behind the board acid aim at making his take off from there.
(2) Luz Long advised Josse to take rest for a few minutes.
(3) Luz Long advised Josse to forget that he was participating in a competition.
(4) Luz Long advised Josse not to accept the Master Race Theory of Hitler.
Man is born, according to H.W. longfellow……….
No sooner did We go out on the road…………. it began to rain heavily.
Ravi said to his friend, “work hard”.
Either Munni or Chunni………lost her pen.
If you lose your passport in a foreiegn country it will effect you badly.
What does the Ganges symbolize, according to Jawaharlal Nehru?
(1) It is a symbol of India’s prosperity.
(2) it is a symbol of India’s age long culture and civilization.
(4) it is a symbol of India’s struggle for freedom.
What damages the crop in lencho’s filled.
What does Lorenzo order when he learns that Portia is on her way to Belmont?
(4) He makes an impassioned plea to the court to intervene on his behalf.
According to Lorenzo plan how we Jessica escape from her father’s house?
(1) She will disguise herself as Lorenzo’s torch bearer and slip out undetected.
(2) She will leave during the night while Shylock is asleep.
(3) She will take her father to a large public auction and get lost in the crowd.
(4) She will fake her own death.
How does Shylock initially describe his demand for a pound of flesh to Bassanio and Antonio?
What reason does Antonio give for being sad in the opening scene of the play?
(1) He stand to lose a fortune in this present business Ventures.
(2) He owes a fantastic sum of money to Shylock.
(3) He gives no reason.
(4) The woman he loves does not return his feelings.
Whom does Bassanio agree to bring with him to Belmont?
I can learn poems more easy than speeches.
(1) why don’t you have your eyes tested?
(2) This notice has been altered.
(3) Boys could not have done all this damage.
(4) Steps had been cut in the ice by previous climbers.
Womenhave always Dressed to……….. figure flaws.
“Would you like to come with us?”they said.
(1) They asked me to go with them.
(2) They asked me if I would like to come with them.
(4) They told me to come with them.
निम्नलिखित ईंधनों में से कौन सा न्यूनतम पर्यावरणीय प्रदूषण उत्पन्न करता है?
| 2019-04-19T12:24:28 |
http://www.examxpress.in/paper-uptet-2016-paper-i-dec/
|
0.999999 |
I am a post-human adapted for permanent life in vacuum and micro-gravity. How might my physiology and biochemistry overcome the following challenges?
Please note I have a strong cultural aversion to augmenting my body in order to help me survive: i.e. enclosing myself in artificially-constructed protective shells, etc.
At this stage, I look similar to an enormous shining pangolin with interlocking plates of keratin-analogue that are alloyed with steel to protect against ionizing radiation. My radically engineered DNA and its unparalleled checksum and repair functions keep the doctor away. I can move these plates to help dissipate heat and - just like a Terran pangolin - curl up in a protective ball to shield my more sensitive parts. I have a reactive outercoating that allows me to alter my pigmentation (which I typically set to polished silver in order to minimise heat loss).
Space is a tough place to live and its very big, so you could say my metabolism has multiple redundancies. I can eat comets and asteroid material, using the spinneret on my tail to swaddle them in cocoons before supping on the extracted organics and minerals. I sift small concentrations of hydrogen from the interplanetary medium for use as a metabolic catalyst, and I'm a radiotroph, with the outer layer of my armoured plates coated in a melanin that helps me capture energetic rays to power my body. Stick me near a magnetic field and I'll even generate power like a dynamo.
I have cat-like eyes, optical amplifiers to better perceive in very weak light. I can arrange sensory organs in my head, four limbs and tail to provide a rough 360 degree scan across the electromagnetic spectrum. My brain can intercept and interpret radio communications, etc.
a. Laser communication I can emit coherent lasers through an adaptation incorporating fluorescent protein from a jellyfish. I use these to communicate with my peers, but I don't know what the sensory receiver looks like. I'm guessing it is a specific point(s) on my body, but what might it be biologically composed of?
b. Celestial mechanics In order to navigate the Titan-Mercury transit properly, I need to be able to sense mass, motions of celestial objects, etc. What kind of biological mechanisms could support such a sensory apparatus?
A) skin - Human skin already has constructs designed to sense light. What you need is to just increase the resolution, response time, and then plug that into a neural framework. But sense you are working with lasers on a cosmic scale you want your receiver to be as big as possible so you don't miss.
B) Technically this is what Cochlea in the Ear is for. Its a ball filled with fluid that gives you your sense of up and down and helps with balance. The fluid is pulled down by gravity.
Granted the human cochlea is way to primitive for what you are asking, but the design principals still work. You just need to increase the neural fidelity so that the tiniest pressure differences orient you towards the greatest gravitational pull. You also need a way to phase out physical movements from interfering with the perception of the sensed field. Again, the human brain already does this to some extent, both consciously and subconsciously.
I don't know what the sensory receiver looks like. I'm guessing it is a specific point(s) on my body, but what might it be biologically composed of?
Well, laser light is just made out of photons like normal light. The only difference being that those photons are coherent, having all the same phase, polarization and direction of propagation.
So, to just detect laser light you don't need anything more sophisticated than an eye, like any spectroscopy lab guy can tell you (together with the pun of not staring into the laser with your remaining good eye). You might just need to be able to withstand higher energy densities.
If you instead want to be able to discriminate between laser and non laser light, you might need some interferometer inside your eyes to sense the phase difference.
I need to be able to sense mass, motions of celestial objects, etc. What kind of biological mechanisms could support such a sensory apparatus?
Here I will go for something cool, since you are post human. If you can detect gravitational waves, you can sense moving mass in the universe. The lower the intensity you can detect, the lower the moving mass you can spot.
If you have an interferometer in your eyes for spotting laser, you can use the same concept to also detect gravitational waves, same as LIGO/VIRGO does.
Not the answer you're looking for? Browse other questions tagged space-colonization genetic-engineering adaptability vacuum or ask your own question.
What does Space exploration look like in a Future with no progress?
How would someone hide an asteroid in the void of space so only they could find it?
What would the perfect human race be like?
Is this atmospheric composition breathable, and what would it look like?
| 2019-04-26T10:15:01 |
https://worldbuilding.stackexchange.com/questions/129647/what-does-the-physiology-and-biochemistry-of-a-vacuum-adapted-post-human-look-li
|
0.69827 |
The ClickView Digital Video Recorder (DVR) includes a scheduler function for use with the supported capture cards. This allows the user to configure the recordings based on 'Time', 'Channel' and 'Date'. The following article details the use of these features.
1. Schedule your recording's 'date', 'start time', and 'stop time' as desired.
2. Select the 'Source' for the scheduled recording. The number of 'Source' options depends on the video capture card used. The Source list may include 'Composite', 'Analogue TV', and 'Digital TV'.
3. Select the Channel for the scheduled recording.
4. Press 'OK' to confirm the selection.
5. Check the scheduled recording, taking note of the source channel which has been selected.
6. Continue to add as many scheduled recordings as required for the various channels. The ClickView Digital Video Recorder will automatically record all scheduled items across the different channels.
| 2019-04-21T18:32:42 |
https://www.clickviewsupport.com/hc/en-us/articles/214580528-PB066-ClickView-Digital-Video-Recorder-Scheduler
|
0.999998 |
Questions about example sentences with, and the definition and usage of "Warming"
What does warming to the subject mean?
What does warming up to her mean?
It could mean to become close to her (maybe as a friend) For example, there is a person that doesn’t seem to like you but then you do favors and act nice towards her. She begins to like you and become more comfortable talking to you/being with you. You are warming up to her.
Please show me example sentences with heart warming.
It is so heart warming to have my wife give me a kiss.
Please show me example sentences with global warming.
Nowadays, global warming is a serious problem.
Please show me example sentences with warmth and warming!! what is the difference in use between these words??! .
Please show me how to pronounce government,difference,global warming.
"Want" usually means "desire". I don't think I've heard it used where it implied "need". "I want to have a cat" will always mean "I desire a cat". It will never mean "A cat is necessary for me". I've also never heard "This car wants washing". First, because a car can't "want". "This car NEEDS washing" is what I would say.
The woman is hoping to be pampered. She is expecting her holiday to be luxurious.
*** Integrated writing task- toefl ibt test *** This is a discussion on the reason of global war... does this sound natural?
*** Integrated writing task- toefl ibt test *** This is a discussion on the CAUSE of global warming. The passage believes that man-made CO2 is the direct cause of global warming, but the lecturer thinks JUST THE OPPOSITE. The passage STATES that global warming is a natural phenomena during earth’s life and THAT it is not due to CO2 release. It adds, “The main source of CO2 on THE atmosphere is carbon sinks, like oceans. This is the direct cause of global warming during THE past 250,000 yrs”. But the lecturer says these ideas are not based on reality from a comprehensive point of view. She SAYS, “THE Isotopic ratio of CO2 from A carbon sink is not equal to that generated BY fossil fuEL combustion. These two CO2 act differently regarding global warming”. She believes that the main cause of global warming during 20th century is man-made CO2. Also the passage thinks that computer simulations ARE UNReliable IN PREDICTING future weather. On the other hand, THE lecturer points OUT THAT scientists HAVE relIED on computer predictions FOR other trends, but not FOR this one - WHY THE INCONSISTENCY?
can someone explain something to dumb people like me ? what is global warming ?
An increase to the earths temperature that affects the nation.
It had been warming but it suddenly got cold today. 最近は暖かったけど今日はまた寒くなった does this sound natural?
It has been warming up (or simply warm) recently, but it suddenly got cold today.
Global warming is a process of a long-term increase in annual average temperature of the Earth's ... does this sound natural?
Global warming is a process of a long-term increase in annual average temperature of the Earth's atmosphere and oceans. The average temperature went up 0.7°C in comparison with the beginning of the industrial revolution. (1850s). Anthropogenic factors, primarily greenhouse gas emissions, such as CO2 and CO4, are the number one cause of global warming. According to climate models the average temperature of the ground surface will rise from 1.1°C to 6.4°C in the 21st century. In some regions, temperature may go down a little. According to many forecasts, the temperature will first drop by a degree in Europe because of the Gulf stream slowing, but later on will increase rapidly. Global warming and sea-level rise are expected to grow over thousands of years, even in the case of greenhouse gas level stabilisation. The rising of sea levels and global warming will result in changing the amount and distribution of precipitation. As a consequence, natural disaster such as floods, droughts, and hurricanes may become more frequent, crop harvest will decline, and many species will go extinct. Climate systems are changing under the influence of both natural internal and external processes. The Greenhouse effect is one of the main causes of change. The Greenhouse effect is a process whereby absorbtion and emittion of infrared radiation by atmospheric gases causes the warming of earth's surface and atmosphere. Pollution sources are animal husbandry, use of fertilisers, coal combustion, car emissions, factories chimneys and other anthropogenic emissions. Landscape change such as deforestation also contributes to the greenhouse effect. Scientists forecast the continuation of global warming, increase of greenhouse gas emissions, growth of average temperature, and sea-level rise. The most probable changing weather patterns are as follows : 1. More heavy precipitation 2. Higher maximum temperatures, more hot summer days and fewer cold winter days in almost all regions of the world, heat waves will become more frequent in the majority of inland areas. 3. Decrease in the range of temperature.
Example sentences with, and the definition and usage of "coexist"
Example sentences with, and the definition and usage of "verve"
Example sentences with, and the definition and usage of "warlike"
Example sentences with, and the definition and usage of "warlock"
Example sentences with, and the definition and usage of "warlord"
Example sentences with, and the definition and usage of "warmd"
Example sentences with, and the definition and usage of "warmduscher"
Example sentences with, and the definition and usage of "warmful"
Example sentences with, and the definition and usage of "warmheart"
Example sentences with, and the definition and usage of "warmhearted"
Example sentences with, and the definition and usage of "warmly"
Example sentences with, and the definition and usage of "warmness"
Example sentences with, and the definition and usage of "warmongering"
Example sentences with, and the definition and usage of "warms"
Example sentences with, and the definition and usage of "warmups"
Example sentences with, and the definition and usage of "warna"
Example sentences with, and the definition and usage of "warnen"
Example sentences with, and the definition and usage of "Warner"
| 2019-04-25T15:00:56 |
https://hinative.com/en-US/dictionaries/warming
|
0.999999 |
This article is about the planet. You may be looking for the species.
Duro (also known as Duros) was the heavily polluted and depopulated homeworld of the Duros species. Located on the Corellian Trade Spine in the Core, the planet itself was mostly abandoned, mainly housing food processing plants. Most of the population lived in one of the twenty orbiting cities. Duro was also home to a large number of orbital shipyards.
Duro prior to the Mandalorian Wars.
Duro was surveyed by the Columi in 100,000 BBY, but the Columi were not impressed and passed it off as a primitive failure, little knowing that the inhabitants of the world, the Duros, would become a space-faring species in the years to come. Under the Infinite Empire, Duro was a slave world, and helped create the Star Forge 30,000 BBY. Following a plague that struck the Rakata, the Duros rebelled in 25,200 BBY. The Duros then formed a hereditary monarchy. It was based at Ranadaast, a city named after Queen Rana Mas Trehalt, which fell into disrepair and was infested by criminals when the monarchy fell into decline. The monarchy declined due to the rise of corruption, and control by a merchant council. Several years after the fall of the monarchy, Duro became a member world of the Herglic Trade Empire, the most powerful government in the Core Worlds region at the time.
Duros were interested in exploration and ventured off their homeworld to explore the galaxy at an early age. Hyperspace cannons linked Duro with Coruscant, Corellia, and other Core Worlds. Having reverse-engineered the hyperdrive independently of the Corellians, Duro entered a Golden Age as a Core Founder 25,053 BBY under Queen Rana Mas Trehalt at the formation of the Galactic Republic. The monarchy was eventually abolished in favor of a collection of ruling corporations. The Duros High House became the government/legislature, and the Chief Representative Officer became the planet's head of state. At the same time, the Herglic Trade Empire fell with the formation of the new galactic government.
The Duros colonized Neimoidia during their earliest wave of expansion, though the Neimoidians were considered a separate culture by 15,000 BBY. The Duros discovered Sneeve in the Kymoodon Era, Adner in 5000 BBY, and Koboth (prior to 219 BBY). The following years the Duros abandoned the planet for space cities, and left food production and mining on the planet to automated systems; however overdevelopment caused mass extinctions and pollution on the world.
3962 BBY, the planet itself was laid to near-total waste during the Mandalorian Wars, when the Mandalorian invasion fleet advancing on the Colonies and Core regions destroyed much of the world using thousands of Basilisk war droids. Years of misery and rebuilding were ahead of the Duros people, merely the latest phase in the long series of Sith Wars. During the New Sith Wars, the Duros colony on Harpori was wiped out by Kaox Krul 1002 BBY.
Circa 220 BBY, during Duro's Industrial Age, the mining city of Tayana reached its height. In 200 BBY, the Tiercam Dam was built.
During the Clone Wars, the Confederacy of Independent Systems launched Operation Durge's Lance in 21 BBY, a coordinated and major campaign that led to the capture of Duro. During the battle, General Grievous ordered an orbital bombardment of the surface. This resulted in the pollutants from the numerous factories which covered Duro's surface to be released radioactively, and the surface being rendered the uninhabitable wasteland that it would remain until being Vongformed years later. The swift success of the campaign sent shockwaves throughout the Galactic Republic, and even neutral Corellia began to tighten its defenses.
Fefze beetles attack Luke Skywalker, Han Solo and Ken.
The Galactic Empire collected what was left of Duro's natural resources, and forced many Duros to move off the planet. In an attempt to stir up conflict between Duro and Corellia, the Empire declared Duro to be part of the Corellian sector. With the Trade Federation nationalized, the disgraced Neimoidians began to reintegrate themselves into mainstream Duros culture. At some point, Darth Vader's secret apprentice, Starkiller, was sent to Duro to dispose of a local despot who had grown too over-ambitious. While the precise details of the incident are unclear, All-Terrain Scout Transports were used against the darksider, earning his unending hatred of the walkers.
The space city Pellezara station, located near the planet in the Duro system, served as a maintenance stop over for freighters en route to Duro.
In 1 BBY, the Rebels tried to free the planet but they were defeated at the Battle of Vnas.
As of 5 ABY, the prisoner Triclops was being held on Duro, courtesy of Grand Moff Bertroff Hissa. Triclops escaped and saved his son Ken from Hissa. Although the New Republic managed to save Duros archaeologists from the Valley of Royalty, the Valley itself was flooded with toxic waste when the Tiercam Dam burst. The New Republic eventually liberated the planet, and put the system back in the Duro Sector. Ecosystem purification teams cleaned up the planet's atmosphere as best they could. 11 ABY, in Operation Shadow Hand, Duro was retaken by the Empire, but eventually freed.
13 ABY, CorDuro Shipping took control of the Pri-Andylan shipyards. 25 ABY, Jedi Tresina Lobi and Thrynni Vae discovered CorDuro Shipping's ties to the Peace Brigade, resulting in the Brigade murdering Thrynni.
The Gateway Dome refugee camp on Duro in 26 ABY.
26 ABY, the New Republic attempted to make Duro habitable and settle refugees there, though their efforts were sabotaged by the Yuuzhan Vong agent Nom Anor. The Yuuzhan Vong conquered Duro and destroyed nineteen of its twenty floating cities. During their occupation, they managed to reverse centuries of pollution in just two years with their biotechnology. It was later raided by the New Republic, and again by the Galactic Alliance.
After the war's end, Duro returned to its lush forested past being restored by the Yuuzhan Vong terraforming. During the Second Galactic Civil War Thrackan Sal-Solo had planned to conquer it to gain access to its baradium and cortosis.
After the Battle of Caamas and the end of the Sith–Imperial War, a Duros Admiral named Gar Stazi led the remnants of the Galactic Alliance Core Fleet against the newly established Galactic Empire. Duro itself was subjugated by the Empire and was subject to a pervasive Imperial presence. An Imperial Academy was established in the planet; in it the most talented Duros taught Imperial pilots and capital ship crewmen.
Eventually, due to millennia of pollution, the only remaining creature on Duro was the nonnative Fefze beetle, an aggressive mutation of the insectoid Fefze, a species recruited to work on Duro during the last decades of the Galactic Republic. Extinct creatures included the cannibal arachnids. When the Yuuzhan Vong arrived on Duro, they introduced many of their own creatures and creations onto the planet.
The PC game Star Wars: Rebellion spells this planet "Duros."
| 2019-04-19T17:01:12 |
https://starwars.fandom.com/wiki/Duro/Legends
|
0.997028 |
A new study shows that wetland regions emitted significantly less methane during glacial times while methane emissions by forest fire activity remained surprisingly constant from glacial to interglacial times.
Using novel isotopic studies, scientists from the European Project for Ice Coring In Antarctica (EPICA) say this identifies the most important processes responsible for changes in natural methane concentrations over the transition from the last ice age into our warm period.
Ice cores are essential for climate research because they represent the only archive which allows direct measurements of atmospheric composition and greenhouse gas concentrations in the past.
In the current study, the scientists present the first glacial/interglacial record of the carbon isotopic composition of methane (δ13CH4) providing essential information on the sources being responsible for the observed CH4 concentration changes.
The well known glacial/interglacial changes in atmospheric methane concentrations are quite drastic. Glacial concentration were on average 350 ppbv (part per billion by volume) and increased to approximately 700 ppbv during the last glacial/interglacial transition superimposed by rapid shifts of about 200 ppbv connected to rapid climate changes. During the last centuries human methane emissions artificially increased CH4 concentrations to approximately 1750 ppbv.
But what caused these substantial changes in natural atmospheric CH4 concentrations prior to the human impact? To answer this question, the scientists developed a new analytical method that now allows to quantify changes in the isotopic ratio of 12CH4 and 13CH4 in ice core samples. This ratio provides insight into the responsible methane sources.
“These studies bring us much closer to a quantitative understanding of what happened with wetlands and methane in the past”, says Dr. Hubertus Fischer from the Alfred-Wegener-Institute for Polar and Marine Research, who is the lead author of the publication and coordinator of the gas studies on the EPICA ice cores. “This is essential to also improve our predictions of how the methane cycle will respond to an increased warming in the future”, he adds.
The study shows, that tropical wetlands emitted substantially less CH4 during glacials; most likely caused by changes in monsoonal precipitation patterns. Together with a reduced atmospheric lifetime, this explains major parts of the glacial CH4 reduction. In addition, boreal methane sources located in wetlands in higher northern latitudes were essentially switched off during the glacial due to the expansion of the northern ice sheets and the very cold temperatures in high northern latitude.
However, these high latitude wetlands were quickly reactivated when rapid climate warming events occurred. Also forest fires emit a considerable amount of CH4, which, however, remained surprisingly constant over time. The isotopic measurements show no signs of CH4 emissions by a destabilization of marine gas hydrate reservoirs when climate was warming.
The current results were published by a team of scientists from Germany, France and Switzerland. As the German partner within EPICA, the Alfred-Wegener-Institute was responsible for the drilling operation of the ice core used for this study. In addition, it specialized on the development of new analytical techniques to measure isotopes in greenhouse gases and the interpretation of changes in biogeochemical cycles in the past. Coordinated by the European Science Foundation (ESF), EPICA is funded by the participating countries and the European Union.
EPICA is one of the core projects of the AWI Research Program “Maritime, Coastal and Polar Systems” in the “Earth and Environment” research section of the Helmholtz-Gemeinschaft. For its outstanding effort and large impact on climate research, EPICA has recently received the Descartes Prize for Transnational, Collaborative Research awarded by the European Commission.
Article: “Changing boreal methane sources and constant biomass burning during the last termination” by Fischer et al.,Nature, April 17, 2008.
| 2019-04-25T10:10:29 |
https://www.science20.com/news_releases/30_000_years_of_methane_sources_show_natural_correlation_with_glacial_periods
|
0.99984 |
The purpose of this board is to allow the development of remote control applications. Using PIC-GSM you can easily monitor and control your home from everywhere with your GSM cell phone. You can switch on and off relays remotely by SMS. You can monitor the status of your GPIOs or toggle them. You can establish a call to the GSM module and speak or listen on the audio channel of the board. You can monitor the temperature from several digital remote temperature sensors. The build-in GSM module has an on-board antenna and it allows you to use it as a GSM or a GPRS meaning that you can also use this module as a GPRS modem connected to your computer. A panic button and alarm inputs can dial your phone or send you a SMS message to report for an alarm.
The PIC-GSM module was working fine but I decided to test how it would behave without the battery. I changed the jumpers as per datasheet but now it doesn't work. What might be the reason?
The battery might be completely out of charge - let it recharge a bit and attempt again. To switch the positions of SMT jumpers you would need additional tools and further experience in electronics. If you dobut your skills do not attempt to change the jumpers. Double-check if all SMT jumpers have been closed (pads soldered) or opened (wire between pads is cut and no soldering residue is present) as per datasheet. IMPORTANT! Always ensure that the three jumpers 4V_E, 4V and 5V_CHG_E jumpers are moved together. If you leave one of those jumpers unchanged and you power the board you would damage the GSM component.
| 2019-04-23T05:00:17 |
https://www.olimex.com/Products/PIC/Development/PIC-GSM/
|
0.999996 |
The path to ending the state budget impasse is becoming increasingly clear. Two plans with competing visions are at the forefront of debate in Harrisburg.
The first option is a tax-and-borrow approach. The plan, which passed the Senate back in July, includes at least $570 million in new taxes and up to $1.3 billion in borrowing. The tax increases would be broad-based, hitting Pennsylvanians in the form of higher utility bills, cell phone bills and online purchases. Borrowing itself would place an additional burden on taxpayers, potentially requiring them to pay more than $2 billion once the interest and principal on the bond is paid off.
Tax hikes and more borrowing would come on the heels of a $650 million tax increase passed last year, along with higher insurance premiums, turnpike fees, liquor prices, and a tax burden that already ranks as one of the most onerous in the country.
The second option involves a plan that respects taxpayers and requires state government to use funds that have accumulated over several years. The proposal—known as the “Taxpayers’ Budget”—calls for the redirection of more than $1.2 billion from 41 shadow budget funds. It also includes nearly $1.2 billion in additional fund transfers, tax changes, and ad revenue.
The Wolf Administration has panned the plan, asserting the fund transfers would force cutbacks in various services. But as we made clear last week, lawmakers want to transfer surplus funds from reserve accounts that have grown over the past several years. Every single penny promised to agencies for the 2017-18 fiscal year remains in place. Additionally, even after the transfers, many funds will still have surplus funds in their individual accounts.
Yet, the bureaucracy in Harrisburg continues to send out dire warnings about the fund transfers without demonstrating how the transfers will immediately and irreparably harm services provided—many of which amount to special giveaways.
It's important to note the Taxpayers’ Budget calls for using approximately 13-14 percent of reserve funds to balance the budget. That means more than $8 billion would be available to pay for emergencies or contractual obligations.
Lawmakers also have the option of raising additional revenue through liquor privatization and gambling expansion.
What's the major takeaway? There are plenty of ways to balance the budget without raising taxes.
| 2019-04-19T10:15:24 |
https://www.commonwealthfoundation.org/policyblog/detail/lawmakers-choice-take-from-taxpayers-or-rein-in-reserves
|
0.999997 |
Can anybody tell me what this text means?
This is about OCD (obsessive compulsive disorder) by the way.
@starsdiedforyou Ohhhhh Thank you very much now it makes sense to me! haha Thanks!!!
@naa “In an effort to prevent one of her fears coming true...” She has a belief that her repetitive actions stop bad things happening. So (for example) if she counts the cars she won’t get run over. However, it is temporary because the irrational fear will return. She has connected her repetitive actions with external events which are not related.
She has a belief that her repetitive actions stop bad things happening.
So (for example) if she counts the cars she won’t get run over.
However, it is temporary because the irrational fear will return.
She has connected her repetitive actions with external events which are not related.
| 2019-04-19T15:29:58 |
https://hinative.com/en-US/questions/12594796
|
0.999429 |
WASHINGTON -- A key House committee chairman said his panel would take a "more thoughtful" approach to legislation allowing states to force retailers to collect sales taxes on Internet purchases.
His comments signaled that momentum from Monday's easy passage of the bill in the Senate won't lead to quick House action on the controversial issue.
Rep. Bob Goodlatte (R-Va.), who heads the House Judiciary Committee, said he had problems with the proposed Marketplace Fairness Act, and criticized the Senate for bypassing the committee process by taking the bill straight to the floor.
Supporters of the legislation, who have been working years on the issue, said the bill was stalled in the Senate Finance Committee and that it was time to address the problem faced by conventional retailers. Those retailers have argued that they must compete against online merchants who do not have to charge sales tax.
Senate Finance Committee Chairman Max Baucus (D-Mont.) comes from a state that does not have a sales tax, and he opposed the bill.
Quiz: How much do you know about Internet sales taxes?
Of the 10 senators from such states -- including Delaware, New Hampshire and Oregon, as well as Alaska, which has no state sales tax but allows local sales levies -- seven voted against the bill.
Those senators argued it was unfair to force retailers in their states to collect sales taxes for other states, but their opposition couldn't stop the bill. The Senate voted 69-24 to approve it, as 46 Democrats joined with 21 Republicans and two independents.
House supporters of the legislation hoped the strong bipartisan support would push the bill forward in their chamber. They have been trying to counter arguments by some anti-tax advocates that the legislation amounts to a tax increase.
Consumers in states with sales tax are obligated to pay it on their online purchases if the retailer doesn't collect it, but only about 1% of people do.
"Let's not wait until more businesses go under because of unfair online competition," said Rep. Jackie Speier (D-Hillsborough), one of the main sponsors of the bipartisan House bill. "This is not a new tax, the technology is there, and it is time for the House to act."
Rep. Steve Womack (R-Ark.), the lead sponsor of the bill, said the House needed to pass it to save "local retail business."
But House Republican leaders have deferred to Goodlatte -- and he's got problems with the legislation.
Goodlatte said he understands the concerns of conventional retailers that they are competing with Internet sellers who do not have to collect sales tax.
But he said the bill's process for collecting sales tax for hundreds of different jurisdictions and then remitting payments needed to be simplified.
Goodlatte has also echoed concerns of some conservative opponents of the bill that it could set a precedent for states to collect other taxes, such as income or business-use levies, from people or companies outside their borders.
"I am open to considering legislation concerning this topic, but these issues, along with others, would certainly have to be addressed," Goodlatte said in a statement after the Senate vote.
"The committee will also look at alternatives that could enable states to collect sales tax revenues without opening the door to aggressive state action against out-of-state companies," he said.
The House Judiciary Committee is likely to hold hearings and consider amendments, said Brian Bieron, senior director of global public policy at EBay Inc., which opposes the legislation in its current form.
The legislation would override a 1992 Supreme Court decision that prevented states from collecting sales taxes from companies with no in-state presence.
California and some other states have begun expanding the definition of an in-state presence so they could begin collecting sales tax from some large online retailers, notably Amazon.com Inc.
The Senate bill would allow states to force larger retailers with no in-state presence to collect sales tax. Companies would be exempt if they have less than $1 million annually in out-of-state online and other so-called remote sales, such as catalog and mail order business.
EBay wants the small-business exemption increased to those with up to $10 million in annual remote sales or to those with fewer than 50 employees.
"The key thing is not the speed but the thoroughness and the openness of the process," Bieron said of House consideration of the legislation. "People are going to have opportunities to raise questions, raise issues."
| 2019-04-18T15:05:59 |
https://www.latimes.com/business/la-fi-mo-internet-sales-tax-house-goodlatte-20130507-story.html
|
0.999743 |
An anonymous reader quotes a report from The Guardian: Workers at the wrecked Fukushima Daiichi nuclear power plant have begun removing fuel rods from a storage pool near one of the three reactors that suffered meltdowns eight years ago. The plant's operator, Tokyo Electric Power (Tepco) said on Monday that work had begun to remove the first of 566 used and unused fuel assemblies in reactor building No 3. The fuel rods stored in unit No 3's cooling pool were not damaged in the 2011 disaster, when a powerful earthquake and tsunami knocked out Fukushima Daiichi's backup power supply and triggered the world's worst nuclear accident since Chernobyl, 25 years earlier. Tepco said the operation to remove the fuel rods, which are in uncovered pools, would take two years, adding that transferring them to safer ground would better protect them in the event of another catastrophic earthquake. Workers are remotely operating a crane to raise the fuel from a storage rack in the pool and place it into a protective cask. The whole process occurs underwater to prevent radiation leaks. The utility plans to repeat the procedure in the two other reactors that suffered meltdowns.
9to5Mac has learned of several new features expected to be included in iOS 13. From the report: Dark Mode: There will be a system-wide Dark Mode that can be enabled in Settings, including a high contrast version, similar to what's already available on macOS. Speaking of macOS, iPad apps that run on the Mac using Marzipan will finally take advantage of the Dark Mode support on both systems. Multitasking: There are many changes coming to iPad with iOS 13, including the ability for apps to have multiple windows. Each window will also be able to contain sheets that are initially attached to a portion of the screen, but can be detached with a drag gesture, becoming a card that can be moved around freely, similar to what an open-source project called "PanelKit" could do. These cards can also be stacked on top of each other, and use a depth effect to indicate which cards are on top and which are on the bottom. Cards can be flung away to dismiss them. Undo gesture: With iOS 13, Apple is introducing a new standard undo gesture for text input on the iPad. The gesture starts as a three-finger tap on the keyboard area, sliding left and right allows the user to undo and redo actions interactively. Safari improvements: Safari on iOS 13 for the iPad will automatically ask for a desktop version of websites when necessary, preventing a common issue where websites will render their iPhone version even when running on an iPad with a big screen. YouTube is notorious for this behavior, forcing users to rely on a 'Request Desktop Site' button. Font management: Font management is getting a major upgrade on iOS 13. It will not be necessary to install a profile to get new fonts into the system anymore. Instead there will be a new font management panel in Settings. A new standard font picker component will be available for developers and the system will notify the user when they open a document that has missing fonts. Smarter Mail: The upgraded Mail app will be able to organize messages into categories such as marketing, purchases, travel, "not important" and more, with the categories being searchable. Users will also be able to add messages to a "read later" queue similar to third-party email apps. Improved multiple item selection: The focus on productivity on iOS continues with the inclusion of new gestures to allow for the selection of multiple items in table views and collection views, which make up for most of the user interfaces found in apps that list large amounts of data. Users will be able to drag with multiple fingers on a list or collection of items to draw a selection, similar to clicking and dragging in Finder on the Mac. New Volume HUD and other changes: Other features to come with iOS 13 include a redesigned Reminders app, which is also coming to the Mac, a new volume HUD, better "Hey Siri" rejection for common mistaken noises such as laughter and crying babies, better multilingual support for keyboards and dictation, and expanded in-app printing controls. Apple is expected to officially unveil the next major iPhone and iPad OS at its annual Worldwide Developers Conference on June 3rd.
Researchers at the University of Toronto have found that cavities may be collateral damage from an overzealous immune system. New Atlas reports: Traditionally, bacteria have taken most of the blame for cavities and tooth decay. The bugs cling to your teeth as plaque and produce acid as waste, which dissolves tooth enamel, dentin and even filling material. But the new study suggests the story is more complicated than that. Oral immune cells called neutrophils are dispatched by the body in response to invading bacteria -- but the researchers found that they might be a little careless in the battle. On their own, neutrophils can't damage teeth but the problems arise after acids from bacteria demineralize them. Once weakened, enzymes released by the neutrophils could wreak havoc on other tooth substances. Damage was found to appear in a matter of hours, and worse still, it also seems to apply to tooth-colored fillings, which may explain why they tend to fail within five to seven years. The silver lining of the discovery is that it could lead to new types of treatment, or new standards for testing materials that are to be used in fillings. The research was published in the journal Acta Biomateralia.
John Timmer from Ars Technica got a chance to take a look at Trek's new bicycle helmet that they claim offers "the first major change in helmet technology in years," and is backed up with peer-reviewed science. Here's an excerpt from Timmer's report: WaveCel is the product of orthopedic surgeon Steve Madey and a biomedical engineer named Michael Bottlang. The two had been working on a variety of ideas related to medical issues and protective gear, funded in part by federal grant money. When considering the idea of a lightweight material that could evenly distribute forces, Bottlang told Ars that they first focused on a honeycomb pattern. But they found that it was actually too robust -- the honeycomb wouldn't collapse until a lot of force had been applied, and then it would fail suddenly. The design they eventually developed has a shape that allows flexing almost immediately when force is applied. "It starts to glide right away," Bottlang said. The manufacturing technique creates a clear point of failure that allows more extensive flexing once a certain level of force is exceeded -- part of the structure will fold over rather than experiencing a complete failure. Then, once folded, the polymer it's made of will allow neighboring cells to glide over each other. This provides some resistance even after the structure has collapsed. For the helmet, a patch of this material is attached to the inside of a more traditional EPS helmet, which provides impact resistance. But the WaveCel mesh is allowed to float within the helmet and can absorb much of the force of off-axis impacts. The thin strips of soft material that cushion the helmet where it rests on the head (also found in more traditional helmets) are attached directly to the WaveCel mesh. It looks more uncomfortable than it is. Madey, the orthopedic surgeon, said they've done tests that show that, even if placed directly on the skin, the WaveCel mesh wouldn't break the skin under most impact forces. How does their new helmet work? According to a paper authored by Bottlang and Madey, helmets including the material reduced rotational acceleration from impacts by 73 percent compared to a normal helmet. A slip pad within a normal helmet (MIPS) only dropped acceleration by 22 percent, which seems like a substantial difference.
An anonymous reader quotes a report from The Verge: SpaceX successfully landed the center core of its Falcon Heavy rocket on a drone ship last week, but the vehicle accidentally fell into the ocean while in transit to the Florida coast. The company blamed the loss on choppy seas. "Over the weekend, due to rough sea conditions, SpaceX's recovery team was unable to secure the center core booster for its return trip to Port Canaveral," SpaceX said in a statement to The Verge. "As conditions worsened with eight to ten foot swells, the booster began to shift and ultimately was unable to remain upright. While we had hoped to bring the booster back intact, the safety of our team always takes precedence. We do not expect future missions to be impacted." SpaceX does have ways to secure the rockets it lands in the ocean, including a robot known as the "octagrabber" that latches on to the base of the boosters. But because the center core connects to two side boosters, it has a different design than a normal Falcon 9 booster. So the octagrabber cannot hold on to it in the same way. The center core is a modified Falcon 9 booster -- one of three that make up the Falcon Heavy rocket -- which flew last week during the second flight of the Falcon Heavy. "Following takeoff, all three cores of the rocket successfully landed back on Earth: the two outer cores touched down on dual concrete landing pads at the Cape while the center core touched down on the company's drone ship named Of Course I Still Love You in the Atlantic," reports The Verge.
| 2019-04-24T11:57:34 |
http://www.netserv.is/aggregator/categories/1?page=13
|
0.998966 |
Is there a way to archive old notebooks? Preferably they should still be searchable.
I would like to be able to keep some old notes but 'hide' them so they aren't included in searches, etc. Is there a way to do that? I have them all organized in notebooks under a specific stack that starts with "z ..." so I don't have to see them when I'm looking at my current notebooks. However, whenever I do searches or sort notes, they come up. There are a lot of them so I don't think I can export and then re-import when needed without going over the monthly data limit. Thank you in advance for any suggestions!
When I am done with a project (like really done, not the false done before the other shoe drops) I would like to retain the notebook for future refrence, but not crowd my active list with old stuff. Can I create an "Archive" notebook (local only, not synched) and drop in other notebooks?
Hello! I am so unbelievably frustrated! I recently got a Sky Wifi Smartpen and it is neccesary to use EN to support it. I am a student, and after my class I archived my notebook and now I can't upload or synch any notes online? Is my notebook archived for good and I can never add anything to it ever again? Ive read I can't get old information back onto the pen but can I add new information to these notebooks? I connect to wifi, I go to archive it and it says I have already used this paper product and the archive fails. PLEASE BE POLITE!
I have about 2,000 notes in Evernote, and many of them are for old consulting clients. I don't want to lose the notes entirely, but I also don't need them cluttering up my screen. Is there a way to make notes Hidden? Or a way that I can Archive some notes to a new file (like one does with Outlook) thereby preserving them for future reference but getting them out of my main file? Surely others have dealt with this before. What is the best practice for handling this? Or should I create another account and move the old notes over there?
(Archived) Feature Request: Tag Suggestions from webmaster, delicious, KGen extractor and their synonyms. No need for brainstorming!
Your tagging system is crucial and here is how Evernote could improve it: Evernote must implement something fully featured like this:http://handytag.elitwork.com/home.htmlhttps://addons.mozilla.org/en-US/firefox/addon/handytag Explanation from me (and the Firefox add-on author):It makes bookmark tagging easier, by simply providing a complete set of most relevant keywords. These keywords are retrieved from many different sources :the tags you already gave to one of your actual bookmarks.the keywords the webmaster gave to his webpage.the tags given by del.icio.us users for the same page.the tags the keyword extractor found using KGen (https://addons.mozilla.org/en-US/firefox/addon/4788).and finally synonyms for the above found keywords.You no longer would need to make a brainstorming session each time you want to add a new bookmark. It will do it for you! Please forward this feature request to Evernote. If this becomes part of Evernote I will move my complete bookmarking system from a certain competitor over to Evernote and become a Premium member. I'd be happy about a short reply. Thanks!
mac (Archived) restore data without archive file?
| 2019-04-23T00:26:20 |
https://discussion.evernote.com/tags/archive/?_nodeSelectName=calendar_event_node&_noJs=1
|
0.999997 |
What is the key to being successful financially? Is it earning well into the six figures, complete with a big-ticket bonus each year? Is it staying debt-free? Or is it investing successfully and eventually hitting it big on the stock market? The answer may surprise you.
Managing your money successfully is a major factor in being financially successful. Consider this: there are people who worked at a mid-level job their entire lives, saved a good portion of their income, dutifully contributed to their retirement and were able to retire as millionaires. Concurrently, there are people who work high-paying jobs with massive bonuses who end up with nothing saved and no way to retire. What's the difference? One managed their money well; the other did not. The key to being financially successful is not how much money you make, but how well you manage it.
Read on to learn how to manage your money successfully.
Managing your money successfully start with your income. It sounds simple, but the key to success in this regard is to spend less than you earn. If you can do that, you can begin to build success through saving and investing your money. However, that is not always as easy as it sounds, and the key to doing it is budgeting successfully. A budget is more than a list of categories and amounts for your money. It is your monthly guide and your plan to help you reach your goals.
Set the amount you'll spend in each area, you prioritize your spending and saving based on your goals and circumstances, and finally, you track your spending. This lets you know when you need to stop spending in a certain category, and where you have wiggle room or extra money that can be allocated elsewhere. Once you get control of your budget, you will be on your way to managing your money successfully.
The best way to manage your debt is to eliminate as much of it as possible. The first step is to stop using your credit cards, and then work to pay them off. Some debt is harder to avoid, such as student loans or buying a car. For example, you may not be able to afford a car, but you need one to get to work each day. Therefore, it will benefit you to shop around for the best deal on a car and the best interest rates, so you can more quickly pay off that debt. The same goes for student loans. Try refinancing to get a better interest rate, or pay more toward your loans each month so more of your money goes to the principal balance rather than interest.
Another area where debt may be unavoidable is when buying a home. But that doesn't necessarily mean it's a bad investment. Look for a home that you can afford, that's in a good neighborhood, and will appreciate in value. Remember, your mortgage payment should be no more than 30% of your income.
They say you should always pay yourself first, and that applies to building your savings and investment portfolio. There are two different types of savings: your liquid savings and your investments.
We'll start with investments. Investing is a beneficial way to save because it actually earns you money and increases your wealth. If you do not know a lot about the stock market and investing, then it can be really helpful to have a financial advisor help you manage your investments.
Here are a few tips: Mutual funds help you naturally diversify your portfolio, but you need to look for ones with reasonable fees and a good rate of return. Make sure that you are diversifying your portfolio, meaning you should never have the majority of your investments in one single stock. You should also invest in different types of investment types as well. You do not want all of your investments to be in tech or any area in case that sector of the market drops.
Investments may be harder to get to during an emergency, and you may not want to be forced to cash them in if the market down when you need them. That's where a liquid savings account comes in. Your savings account should contain your emergency fund and should be easily accessible in case of an emergency. You want to be able to access your savings fairly quickly if needed, but you also want to earn the best possible interest rate.
Some people will simply store it in a bank savings account, while others will use a money market fund, savings bond, or a CD. This money is not really an investment, but more of a protection for the emergencies. Think of it as your insurance policy, while the real wealth-building occurs in your investments.
How you manage your money once you're married may change. Read our tips here.
Try these five tools for managing your debt.
Want to invest but not sure how? Try these investing tips for beginners.
Create a budget that takes into account your income and necessities.
| 2019-04-20T08:37:33 |
https://www.thebalance.com/how-can-i-manage-my-money-2385524
|
0.91984 |
WHAT DOES THE MARS 2020 MISSION AIM TO ACHIEVE?
The Mars 2020 mission is part of NASA's Mars Exploration Program, a long-term effort of robotic exploration of the red planet.
NASA hopes the mission will help to answer key questions about the potential for life on Mars.
The mission also provides opportunities to gather knowledge and demonstrate technologies that address the challenges of future human expeditions to Mars, including producing oxygen from the Martian atmosphere, and identifying water.
| 2019-04-18T17:27:53 |
https://www.dailymail.co.uk/sciencetech/fb-5400841/WHAT-DOES-MARS-2020-MISSION-AIM-ACHIEVE.html
|
0.910231 |
Three methods were developed for the application of stoichiometry-based network analysis approaches including elementary mode analysis to the study of mass and energy flows in microbial communities. Each has distinct advantages and disadvantages suitable for analyzing systems with different degrees of complexity and a priori knowledge. These approaches were tested and compared using data from the thermophilic, phototrophic mat communities from Octopus and Mushroom Springs in Yellowstone National Park (USA). The models were based on three distinct microbial guilds: oxygenic phototrophs, filamentous anoxygenic phototrophs, and sulfate-reducing bacteria. Two phases, day and night, were modeled to account for differences in the sources of mass and energy and the routes available for their exchange.
The in silico models were used to explore fundamental questions in ecology including the prediction of and explanation for measured relative abundances of primary producers in the mat, theoretical tradeoffs between overall productivity and the generation of toxic by-products, and the relative robustness of various guild interactions.
The three modeling approaches represent a flexible toolbox for creating cellular metabolic networks to study microbial communities on scales ranging from cells to ecosystems. A comparison of the three methods highlights considerations for selecting the one most appropriate for a given microbial system. For instance, communities represented only by metagenomic data can be modeled using the pooled method which analyzes a community's total metabolic potential without attempting to partition enzymes to different organisms. Systems with extensive a priori information on microbial guilds can be represented using the compartmentalized technique, employing distinct control volumes to separate guild-appropriate enzymes and metabolites. If the complexity of a compartmentalized network creates an unacceptable computational burden, the nested analysis approach permits greater scalability at the cost of more user intervention through multiple rounds of pathway analysis.
Complex microbial communities drive the Earth's biogeochemical cycles . In spite of their importance, the biochemical interactions within these communities are not yet well understood, nor have in silico methodologies for studying them matured. An improved understanding of how natural microbial communities mediate biogeochemical cycles will augment predictions of how these processes respond to disturbances from climate change to anthropogenic chemical deposition. Improved understanding may also provide a rational basis for using microbial consortia to produce biofuels and biomaterials from renewable resources .
Microbial communities described in terms of their environmental chemistry, ecophysiology, and phylogenetic diversity can be used as a foundation to develop, test, and compare in silico tools for analyzing community interactions. The phototrophic microbial mats of Octopus and Mushroom Springs of Yellowstone National Park (Wyoming, USA) represent an ideal test case due to the extensive available data [e.g. [3–9], and numerous other references found throughout this study]. These alkaline siliceous hot spring mats (50-74°C) are inhabited predominantly by unicellular cyanobacteria related to Synechococcus spp. and filamentous anoxygenic phototrophs (FAP) related to Chloroflexus and Roseiflexus spp. The community also contains sulfate-reducing bacteria (SRB) and other prokaryotes sustained by the primary productivity of the photosynthetic bacteria .
Previous studies of the mat have revealed diel (day-night) metabolic variation in various community members, driving shifts in the concentration and fate of dissolved metabolites [5, 6, 10–12]. Organic metabolites including glycolate and other acids, along with hydrogen, form the basis for mass and energy exchanges between community members (Figure 1A). For example, during the day, the photosynthetic cyanobacteria consume CO2 and produce O2 as a by-product of photosynthesis. High levels of O2 relative to CO2 promote oxygen competition at the ribulose-1,5-bisphosphate carboxylase/oxygenase (rubisco) active site, leading to the production of glycolate [9, 13, 14]. Other community members, including the photoheterotrophic FAP, can use glycolate as a carbon and energy source . The cyanobacteria can also store excess photosynthate as polyglucose . This carbon and energy storage material is fermented at night to organic acids (Figure 1B) . FAP can incorporate fermentation products photoheterotrophically [12, 15, 17], while SRB appear to oxidize some of these products under anaerobic conditions [6, 15, 18].
Guild interactions. Modeled consortium interactions are shown for (A) daylight and (B) nighttime simulations. Abbreviations: EtOH, ethanol; hv, photosynthetically-available photons; PG, polyglucose; PHB, polyhydroxybutyrate; syn, cyanobacteria (i.e. Synechococcus).
Molecular level metabolic models were constructed to represent the central metabolism of three functional guilds thought to be important to material and energy flows through this community . A guild is a group of species that exploit the same class of environmental resources in a similar manner . Oxygenic photoautotrophs related to Synechococcus spp. were selected to represent the mat's primary source of fixed carbon and nitrogen. FAP from the family Chloroflexaceae, an important structural component of the mat community, were incorporated as metabolically versatile (photo)heterotrophs, capable of capturing light energy as phosphodiester bonds (cyclic photophosphorylation), but requiring externally supplied reducing equivalents. An SRB guild was included as a consumer of the metabolites produced by photosynthetic guilds. Other sub-dominant organisms (e.g. aerobic heterotrophic bacteria and methanogenic archaea) present in the mat were not included in this first in silico representation of the system (see methods).
Metabolic networks were studied using the stoichiometry-based network analysis approach known as elementary mode analysis (EMA) [20–22]. EMA evaluates a metabolic network by decomposing the system into a complete set of the simplest biochemical pathways satisfying steady state conservation constraints with enzymatic fluxes occurring in physiologically reasonable directions (i.e. unidirectional enzyme-catalyzed reactions are not permitted to run backwards). These explicitly defined pathways are known as elementary modes and are frequently represented as vectors containing relative rates for every reaction in the metabolic network. The approach permits the description of any sustainable physiological behavior as a linear combination of the simplest biologically meaningful metabolic pathways [23, 24].
Metabolic network modeling methodologies have been used extensively to analyze single organisms and monoculture systems [e.g. [25, 26]]. An alternative stoichiometry-based technique, flux balance analysis, has been applied to a two-species laboratory microbial system . That study used a compartmentalized approach analogous to strategies for modeling metabolic fluxes between organelles in eukaryotic organisms [28, 29]. Topological approaches to metabolic network analysis have also been used at tree of life scales to analyze evolution of metabolism and the interactions of metabolic networks among species and with the environment [30–34]. However, these solely topological studies implicitly assume all genome encoded genes are simultaneously active with no consideration of critical reaction quantitative aspects like carbon and electron conservation relationships.
In the present study, novel methods for handling multispecies EMA are developed and applied to a microbial consortium representing a well-studied thermophilic community from Yellowstone National Park (YNP). To the best knowledge of the authors, this is the first time a stoichiometry-based network analysis approach has been applied to a natural, in situ microbial community and the first time these approaches have been used to analyze microbial biofilms/mats. Mass and energy fluxes were investigated using three different approaches each representing different levels of detail and computational complexity. The three approaches provide a general set of tools for studying microbial communities that can be selected based on the levels of available knowledge and community complexity. Each approach has advantages and disadvantages which are discussed along with examples illustrating how these methods can be used to gain insight into basic community metabolic processes. In particular, our work demonstrates a trade-off between knowledge of the distribution of metabolic capabilities between guilds in a consortium and the ability to conservatively predict and interpret limits for that community's behavior.
The primary aim of this theoretical study was to develop, analyze, and compare different in silico approaches for molecular-level analyses of mass and energy flux through a microbial community. The following section details three distinct approaches, using the Yellowstone thermophilic phototrophic mat system as a test case. Advantages and disadvantages of each technique are highlighted and discussions detailing which approach is most appropriate for a given set of microbial data are provided. For each of the three methodologies discussed, both day and night scenarios were considered (Figure 1). All metabolic models, including an explicit listing of the reactions and metabolites, can be found in the additional file 1: Supplement.
A compartmentalized model was constructed in which each of the three guilds was a distinct compartment and exchangeable metabolites were transferred through a fourth compartment representing the extracellular environment (Figure 2A). This strategy has been used in the past for spatially and chemically segregated eukaryotic and microbial systems [27–29]. The approach was implemented by assigning reactions and metabolites to a network representing each guild (suffixes on metabolite identifiers prevented sharing of compounds common to the metabolism of multiple guilds), while explicit transport reactions accounted for the exchange of metabolites between guild members and the extracellular space. A short illustration of the separation mechanism is provided in additional file 1: Supplement.
Multi-species modeling approaches. Conceptual depiction of the (A) compartmentalized, (B) pooled, and (C) nested consortium modeling approaches. Dashed red lines indicate system boundaries while thin solid lines represent guild boundaries.
The compartmentalized modeling approach has the advantage of conceptual tractability. Dividing the community into guild-level compartments linked by transferred metabolites (e.g. glycolate) is an intuitive way to represent interactions within a consortium. It is also an ideal method for understanding which guild performs a particular metabolic transformation. For example, the fraction of total biomass (carbon moles) or total maintenance ATP (used to account for energy-dependent cellular processes other than growth) produced by each guild is clear from the output.
One drawback to the compartmentalized approach is the sheer quantity of unique interactions possible. The size of the resulting network can lead to a 'combinatorial explosion' of elementary modes . To address this limitation, the models for each guild member were constructed to capture the necessary metabolic capabilities while maintaining computational tractability (see methods section). A second drawback of the compartmentalized method is the requirement for significant a priori information or assumptions, as enzymes must be assigned to each separate guild. The Octopus and Mushroom Spring phototrophic mats are well-studied, permitting the use of compartmentalized results as a benchmark for the other two methodologies.
The second approach, referred to as the pooled reactions consortium analysis, treated the consortium as a single entity (Figure 2B). All metabolic reactions and metabolites from the three guilds were combined into a single compartment. Reactions catalyzed by more than one guild were only considered once. The method captures the metabolic constraints of the overall matter and energy transformations without the need for detailed knowledge of every organism in the community.
The pooled reactions approach is ideally suited for investigating the metabolic potential of a community based solely on metagenomic data because the assignment of each reaction to a constituent guild is unnecessary. The technique is quite flexible and can be scaled to different levels of detail. An additional advantage of the pooled reactions approach is the reduction of computational burden. With these advantages, the method is uniquely suited for initial and exploratory analyses of diverse or poorly understood communities. For scientists accustomed to well-documented laboratory microbes, this approach may seem coarse however, the vast majority of organisms on the planet have not been cultured in the laboratory, much less isolated to purity . This method represents a flexible starting point for analyzing such systems.
A weakness in the pooled reactions approach is that model output does not specify which guilds employ a particular enzyme or produce biomass and maintenance ATP. Instead, the results describe potential performance of the entire consortium. The method also neglects the logistics associated with transferring metabolites between organisms, including conversion of the given metabolite into one for which transporters are available. The approach is presented here using the well characterized cyanobacterial mats to highlight strengths and drawbacks to the method.
The third approach, termed the nested consortium analysis or EMA 'squared', uses successive rounds of EMA to analyze potential interactions within a consortium (Figure 2C). The first round of EMA is applied to each guild model in isolation. The output data are mined for ecologically relevant elementary modes, such as efficient biomass production from sunlight or efficient ATP generation from lactate and sulfate. The selected elementary modes are then compiled and used as input reactions for a second round of EMA to examine the potential for interactions between guilds. Conceptually, the first round of EMA provides guild-level stoichiometries relating substrates to products. These stoichiometries can be further processed in the same way that traditional EMA uses enzymatic stoichiometries to consider inter- and intraguild interactions. An example of the process is provided in additional file 1: Supplement.
The first round model output was explored for metabolic strategies representing 'selfish' and 'altruistic' operation of each guild. Selfish selection criteria were based on the efficient production of biomass and maintenance ATP. Altruistic selection criteria were based on efficient production of metabolites that could be consumed by other guilds, including for instance H2, NH3, acetate, and glycolate. Efficiency was evaluated by normalizing the amount of production by the required input (e.g. amount of glycolate or sunlight). A list of the ecological criteria considered can be found in additional file 1: Supplement.
Stable numerical implementation of the nested approach using CellNetAnalyzer v 9.0 required some heuristics. Elementary modes from the first round of analysis were normalized to vectors of unit length prior to the second round of EMA; without this step, errors occurred during calculation of elementary modes; specifically, the attempted inversion of matrices that are very nearly singular. Normalized coefficients were rounded to two decimal places to prevent similar problems. When rounding converted an otherwise non-zero coefficient to zero, the entire elementary mode was scaled so the smallest coefficient was equal to 0.01. The MATLAB script used for normalizing and rounding intermediate results is provided in additional file 1: Supplement. This step was necessary although it introduced a modicum of error. Error analysis is included in the following sections. These steps are not necessary if EMA is performed using the new bit pattern tree algorithm which is numerically more stable .
The nested pathway approach allows for the analysis of more complex metabolic networks and larger communities than the compartmentalized approach because each guild is initially considered separately. It also has the advantage of retaining guild-specific output lost in the pooled technique. The manual selection of specific modes permits a guided community analysis based on ecological strategies of interest for each guild; results can then be re-examined in light of the selection criteria used, relating community stoichiometry and guild strategy. The ability to assemble multiple sets of first-round output and concatenate the results allows the system to expand freely. The nested approach is unrelated to the decomposition approach for the analysis of large cellular networks with EMA , but is well-suited for reconnecting the output of the resulting subnetworks tractably.
The nested pathway analysis approach has the disadvantage of requiring two rounds of processing. In addition, manual selection of ecologically interesting modes from individual models requires a priori knowledge and can significantly influence the solution space. Finally, intermediate processing introduces some rounding error.
Two test cases were analyzed to compare the three approaches in terms of their ability to describe and explain the flows of carbon, nitrogen, and energy through the mat community. The first case study considered daylight production of biomass fueled by solar energy, while the second case study considered the fermentation of stored polyglucose to drive nitrogen fixation during the night. Table 1 highlights key model outputs associated with the different daytime simulations, including the number of modes identified as well as important parameters like maximum yield of biomass or maintenance ATP per photosynthetically-available photon. Table 2 shows similar results for the nighttime simulations.
Elementary mode analysis output summary for comparison of daylight modeling approaches.
a 'Yi/j' is the yield of i on j (i.e. production of i divided by consumption of j). 'X' represents Cmoles of biomass and 'hv' denotes moles of photosynthetically-available photons.
b 'ATP' represents moles of ATP available for cellular maintenance.
c 'syn' designates a parameter associated with the Synechococcus guild. 'FAP' indicates a parameter associated with the filamentous anoxygenic phototroph guild and 'SRB' denotes a parameter associated with sulfate-reducing bacteria guild.
d 'PHB' represents Cmoles of polyhydroxybutyrate (a carbon/electron storage compound associated with the filamentous anoxygenic phototrophs).
Elementary mode analysis output summary for comparison of nighttime modeling approaches.
a 'Yi/j' is the yield of i on j (i.e. production of i divided by consumption of j). 'cypc' denotes moles of nitrogen fixed and stored as cyanophycin, while 'PG' represents Cmoles of polyglucose.
b 'SO4' represents moles of sulfate used as an oxidant.
c 'X' represents Cmoles of biomass.
d 'ATP' represents moles of ATP available for cellular maintenance.
e 'syn' designates a parameter associated with the Synechococcus guild. 'FAP' indicates a parameter associated with the filamentous anoxygenic phototroph guild and 'SRB' denotes a parameter associated with sulfate-reducing bacteria guild.
The relationship between solar energy input and community productivity is a fundamental subject of ecological inquiry, as is the relative abundance of various species. The model community is an example of how metabolic knowledge feeds into these questions, providing a natural point for comparison of the three methods. Oxygenic photosynthesis drives primary productivity in the community by splitting water into reducing equivalents and O2. The reducing equivalents can be used to fix CO2 via the Calvin-Benson-Bassham cycle. The O2, however, can compete with CO2 for the rubisco active site, resulting in the production of glycolate instead of an additional Cmole of reduced carbon (equations 1 and 2). A Cmole is a mole of carbon atoms; one mole of glucose (C6H12O6) is equivalent to six Cmoles of glucose.
Experimental characterization of these fluxes from the Octopus and Mushroom Spring microbial mats has been reported. The O2/CO2 flux ratio at the rubisco active site is estimated to be approximately 0.03-0.07 when dissolved gas concentrations represent mid-afternoon conditions at the springs . The derivation of these estimates from the reported data is demonstrated in additional file 1: Supplement.
The output from the three in silico community analysis approaches was plotted using two physiological variables thought to be important to community interactions: 1) O2 competition at the rubisco enzyme (O2/CO2) and 2) the quantity of biomass (Cmole) synthesized per mole of absorbed photons (Figure 3). Relationships between these properties are examined below as a vehicle for comparison of the three modeling approaches.
Daytime case study output. Daylight biomass-producing elementary modes, plotted for (A) compartmentalized, (B) pooled, and (C) nested simulations in terms of total consortium biomass yield on photons (Cmole/mole) and oxygen competition at the rubisco enzyme (O2 flux relative to CO2 flux). Each point represents a single mode. Solid line segments in (A) connect modes combined to represent ecological scenarios, while dashed lines represent the experimentally-derived limits for oxygen competition at the rubisco enzyme in the mid-afternoon. Arrows denote modes producing cyanobacterial biomass (see text for more details). Points are color-coded by biomass-producing guild (A &C) or SRB-associated reductive acetyl-CoA activity (B).
The compartmentalized approach, wherein guilds interact through a mass-balanced external compartment (see section 1a, Figure 2A, and additional file 1: Supplement), served as a basis for comparison of the community modeling techniques. This method revealed only two elementary modes (out of a total of 74,507 distinct elementary modes) producing cyanobacterial biomass. These modes, highlighted with arrows in Figure 3A, differ based on ATP production strategy. The higher-yielding elementary mode generated ATP by oxidizing reducing equivalents from non-cyclic photophosphorylation, while the lower-yielding elementary mode produced ATP through cyclic photophosphorylation.
All other biomass-producing elementary modes generated FAP, SRB, or a combination of FAP and SRB biomass. At low levels of O2 competition at the rubisco enzyme, Figure 3A illustrates that optimal FAP and SRB biomass synthesis efficiency increases with oxygen competition (positive trend for highest biomass-yielding elementary modes for each increment of oxygen competition). This is due to an increase in glycolate production, shuttling electrons either directly or indirectly to the other guilds. The trend continues until a rubisco O2/CO2 flux ratio near 0.13 is reached. At O2 competition levels greater than about 0.17, O2 interferes with the efficient photon-based production of cellular carbon (i.e. the reducing equivalents liberated with the oxygen are insufficient to cover the costs of processing the additional glycolate).
Theoretical ecological studies have suggested that organisms adapted to a biofilm or mat lifestyle will draw on available resources in a manner which maximizes yields and thus favors the whole community [39, 40]. Combining this theory with the experimentally determined O2/CO2 competition at the rubisco enzyme made it possible to predict relative yields of the cyanobacteria and FAP. Two different cases were considered. The first case (i) examined the ramifications of each guild striving to maximize its own biomass yield and the second case (ii) examined different guilds working in concert to maximize biomass for the entire consortium (see Figure 3A).
Cellular metabolism can be expressed accurately as a linear combination of distinct elementary modes [24, 41]. It is hypothesized that a consortium's metabolic activity can also be described using linear combinations of the elementary modes (with non-negative coefficients). When exactly two elementary modes are considered, the contribution of each mode to the overall metabolic process (represented by a point on the connecting line segment) is fixed by the ratio of any two fluxes for which the constituent modes have non-identical ratios. That relationship is demonstrated in additional file 1: Supplement. For each considered case, a combined metabolism was identified using the ratio of fluxes through rubisco (O2/CO2) derived from experimental data . The derived values were used to bound the solution space (i.e. set of acceptable flux distributions).
This relationship was applied to the highest-yielding biomass-producing modes from the compartmentalized model to predict the relative abundance of cyanobacteria and FAP for a fixed photon input budget (line segments in Figure 3A connect the elementary modes of interest). Modes producing SRB biomass could easily be combined in the same way, but were not among the highest-yielding strategies. The scenario normalized biomass to a per photon basis. The case i scenario predicts ratios of cyanobacteria to FAP biomass production ranging from 2.5:1 to 6.5:1 (cyanobacteria to FAP) based on the experimentally determined oxygen competition at the rubisco enzyme. The case ii scenario predicts ratios from 8.2:1 to 17.7:1 on the same basis. The numerical procedure is demonstrated in additional file 1: Supplement. The analysis control volume consists of the top 1 mm of mat. The control volume growth is exactly equal to the rate at which biomass leaves the control volume. Physically speaking, this represents the new growth burying the old growth.
The ratio of cyanobacterial to FAP biovolume in a Mushroom Spring mat sample was experimentally measured to be 1.6:1 . It is assumed here that biomass and biovolume are related in the same manner for both species. The relative abundances of metagenomic reads suggest biomass ratios for these guilds between about 1.5:1 and 3.5:1 in the top 1 mm of Octopus and Mushroom Spring mats, assuming no bias in the sampling and sequencing pipeline (Klatt et al., in preparation). The reported ratios are closer to the modeled scenario where each individual guild seeks to maximize its own biomass yields rather than the scenario optimizing biomass yield of the entire consortium. A potential explanation for the modest discrepancy between prediction and experimental data is the experimental underestimation of glycolate production due to consumption of radiolabeled photosynthate during incubation. Increasing the flux ratio (O2/CO2) at rubisco (to correspond with higher glycolate production rates) drives the solution space for the case i scenario into stronger agreement with experimental values much more quickly than for the case ii scenario. It should also be noted that in situ environmental systems vary considerably in both space and time, and the flux ratios were derived from measurements at a single condition.
The same variables were plotted for elementary modes identified with the pooled technique, wherein all reactions and metabolites were combined into a single metabolic unit (Figure 3B). The predicted maximum community biomass efficiency was 9.4% higher for the pooled approach than the compartmentalized approach. The cyanobacteria (responsible for the highest-yielding, compartmentalized biomass mode) do not possess a full citric acid cycle; with the pooled treatment, however, the entire cycle becomes available, efficiently catabolizing photosynthate without the usual logistical costs of metabolite transport between guilds. The genome of RS-1, a FAP isolated from a hot spring microbial mat, appears to code for the enzyme missing from OS-A and OS-B', sequenced cyanobacteria from Octopus Spring.
Contrary to the compartmentalized approach, the pooled output suggests that oxygen competition at the rubisco active site does not necessarily affect the biomass yield per photon (note the nearly flat trend of uppermost points in Figure 3B). This represents an uncoupling of heterotrophy and photorespiration, neglecting the role of glycolate as a transfer metabolite. These seemingly contradictory elementary modes all utilized the SRB-associated reductive acetyl-CoA CO2 fixation pathway, and are striped gray in Figure 3B. This pathway is metabolically inexpensive in the pooled approach due to the direct availability of reducing equivalents derived from oxygenic photosynthesis. These enzymes are generally very sensitive to O2, calling the feasibility of these particular elementary modes into question. SRB activity has been reported in the upper 0.5 cm of the mat during day , but the activity may be limited to small anoxic microenvironments and is unlikely to represent a major daytime CO2 fixation strategy.
Filtering out elementary modes using the reductive acetyl-CoA pathway leaves a trend of decreasing optimal biomass per photon yields with increasing O2/CO2 flux ratios at the rubisco enzyme. The trend appears very similar to case ii from the compartmentalized analysis as a result of the pooled approach's treatment of biomass. A non-guild-specific biomass term means that maximizing the yield optimizes production of biomass for the consortium as a whole, the same criteria used to generate the case ii scenario.
This analysis highlights potential challenges associated with the pooled community approach. First, pooled reactions can result in higher predicted yields than compartmentalized reactions because there are no logistical costs associated with inter-guild metabolite transfer. The second challenge is the combination of enzymatic activities that are not compatible. The use of oxygenic photosynthesis to drive extremely oxygen-sensitive enzymes, for instance, seems unreasonable. The pooled analysis generates more appropriate trends if a priori information regarding the oxygen sensitivity of the reductive acetyl-CoA enzymes is applied as a filter. In addition, because enzymatic activities are shared rather than apportioned between individual guilds, maximizing any yield does so for the entire consortium rather than a single guild. This approach is well suited for initial work on systems lacking sufficient a priori data to construct a compartmentalized model. The fact that such data is lacking for the vast majority of all natural ecosystems emphasizes the usefulness of this approach.
The nested approach produces similar results to the compartmentalized method. While this may not be readily apparent from a cursory comparison of Figures 3A and 3C, connecting the extreme upper-left and upper-right points on each subfigure shows that the limits of the solution space (in terms of biomass yield on photons and oxygen competition at the rubisco enzyme) are quite comparable, although the nested approach identifies fewer suboptimal solutions. The biomass yield on photons decreases with increasing O2/CO2 flux ratios at the rubisco enzyme. The initial increasing trend seen in the compartmentalized model consisted of suboptimal solutions not identified as ecologically relevant in the first-round processing. Those metabolic behaviors are all outperformed by linear combinations denoted by the case i line in Figure 1A. The optimal biomass yield on photons is equivalent (within the error introduced by rounding) to the compartmentalized prediction. Combination of modes, as in the case i compartmentalized scenario, identifies very similar ratios of biomass productivity for cyanobacteria to FAP (from 2.0:1 to 5.9:1) at the O2/CO2 flux ratios of 0.03-0.07 suggested by experiment (calculations are demonstrated on the compartmentalized output in additional file 1: Supplement). The total number of elementary modes is lower than the compartmentalized approach (428 vs. 74,507) because the initial round of EMA selects only strategies deemed ecologically competitive or relevant. The first round of processing removes many mathematical solutions that are not necessarily of interest for additional analysis. Caution is required; neglect during first-round selection can prevent discovery of interactions or cause underestimations of metabolic potential.
Any system characterized well enough to construct a compartmentalized model can also be analyzed using the nested approach. The approach is best suited for complex systems that prove computationally difficult to analyze using a compartmentalized simulation. In addition, the nested community analysis approach facilitates interrogation of the competitive strategies employed by individual guilds to produce a compartmentalized modeling result. For instance, the experimentally relevant case i results shown in Figure 3A (biomass production efficiency optimized by guild rather than as a community) require the FAP guild to maximize biomass yield (based on carbon influx) for the strictly heterotrophic co-metabolism of glycolate and acetate. The cyanobacterial guild must concurrently operate a metabolism comprised of three distinct elementary modes: the highest yielding strategies for biomass, glycolate, and acetate production all based on photons. This information is embedded in the criteria used to select the first-round elementary modes.
The nested community analysis approach requires the elementary mode coefficients to be normalized and rounded. Error analysis was performed to gauge the effect of this processing on the results. Each mode was examined for the error in carbon fixed and the amount of absorbed energy. The maximum error introduced was 0.004 Cmoles per mole of photons. The complete error results from both case studies are included in additional file 1: Supplement.
Nitrogen fixation fueled by the fermentation of polyglucose is thought to be an important process in the mat during night and early morning [5, 11]. The presented models only consider nitrogen fixation at night, since the oxygen sensitivity of nitrogenase should render it inoperable during the day when the mat is superoxic. Future modeling efforts will consider the morning separately, acknowledging evidence that suggests a large fraction of nitrogen fixation in situ occurs after sunrise but before O2 production exceeds consumption .
The second case study examined the efficiency of N2 reduction with polyglucose and the corresponding production of five potentially inhibitory or toxic fermentation by-products: formate, acetate, ethanol, lactate, and propionate. Their toxicities have not been explicitly characterized in this system, but their total molar yield on polyglucose represents a convenient 'toxicity parameter' for comparison of modeling approaches. It should be noted, however, that nitrogen fixation and the incidental production and fate of fermentation products are separately questions of ecological interest.
The FAP guild was included in the night-time simulations. They were inactive however, because unlike the SRB, they are incapable of drawing on sulfate as an electron acceptor. While the FAP guild can jettison electrons as hydrogen, this requires energy input from light or an alternative oxidant (oxygen was assumed to be unavailable during the night scenario).
Figure 4A depicts the relationship between the secretion of fermentation products and the efficiency of the metabolic strategy according to the compartmentalized results. The high density of points is an explicit illustration of the system robustness. As the yield of NH3 synthesis increases, more fermentation products are produced. If heterotrophic guilds such as the SRB rapidly metabolize the fermentation products (or those products are removed efficiently through physical transport processes), the high-yielding strategy would likely be the most ecologically competitive strategy for the cyanobacteria. If, alternatively, fermentation products are building up to toxic levels, than the mode associated with lowest by-product loading could be preferred.
Night case study output. Nighttime nitrogen-fixing elementary modes, plotted for (A) compartmentalized, (B) pooled, and (C) nested consortium simulations in terms of total fixed nitrogen yields on polyglucose (Nmole/Cmole) and fermentation byproduct loading (total moles of fermentation by-products produced for each Cmole of polyglucose consumed). Each point represents a single mode.
The pooled results in Figure 4B demonstrate the effects of neglecting constraints on the exchange of metabolites between guilds. The results suggest, in contrast to predictions by the other two models, that nitrogen fixation at night is possible without associated production of potentially inhibitory or toxic fermentation products (all points in Figure 4B located on the y-axis above the origin). Similar to the photo-efficiency test case, the pooled analysis provides a less conservative solution space than the other approaches. Less information is required to build the pooled model, but predicted yields have to be viewed as theoretical upper limits, with the actual consortium probably operating at lower yields. The approach offers a rational, constraint-based approximation to the metabolic potential of a poorly characterized microbial consortium.
The night nested results have nitrogen-fixing efficiencies similar to the compartmentalized approach. The similarity of the shape and size of the solution spaces between nested and compartmentalized techniques (Figures 4A and 4C) indicates that modes of interest were selected appropriately during first-round processing. Considering the corresponding first-round criteria allows mapping of results into ecological strategies for individual guilds. For example, in the elementary mode resulting in the highest yield of fixed nitrogen on polyglucose, the cyanobacteria guild uses two strategies, maximizing production of free NH3 and cyanophycin and co-producing acetate, formate, and ethanol. Three distinct SRB physiologies are employed, maximizing biomass production on each fermentation product individually. Error analysis on the night nested model was performed by comparing error in the carbon balance with polyglucose consumption for every mode. The largest error due to rounding was a reasonable 0.033 Cmoles for every Cmole of polyglucose consumed. The complete error results from both case studies are included in additional file 1: Supplement.
The compartmentalized consortium analysis approach was used to explore additional community properties. Metabolic networks within individual cells are highly branched, forming robust and decidedly redundant systems thought to be resilient to perturbations and disruptions [34, 43–45]. Figure 5 details how the 74,507 unique daytime elementary modes can be divided into 16 different inter-guild mass and energy exchanges. Interactions associated with a large number of elementary modes are more likely to remain functional during environmental perturbations [46, 47].
Daytime compartmentalized guild interaction classes. All 74,507 elementary modes identified in the daylight compartmentalized model, binned into 16 classes of guild interactions. Number of modes belonging to each classification is listed in lower right. Columns from left to right contain interaction classes requiring an increasing number of inter-guild metabolite transfers. Abbreviations: PG, polyglucose; PHB, polyhydroxybutyrate; syn, cyanobacteria (i.e. Synechococcus).
Figure 5 indicates that a large number of metabolic strategies (93.9% of elementary modes revealed by the compartmentalized simulation) exist for the daytime transfer of carbon and electrons from the cyanobacterial guild to the FAP guild with associated FAP production of H2 providing electrons to the SRB guild. In fact, daylight SRB activity in silico is overwhelmingly dependent on such hydrogen (97.1% of compartmentalized elementary modes with non-zero SRB fluxes). There are only 58 unique elementary modes involving the transfer of acetate from the cyanobacteria to SRB which do not involve any FAP activity. The interaction analysis also highlights the importance of the well-connected and functionally versatile FAP guild in the mat community metabolism. Only 74 out of the 74,507 unique elementary modes (0.1%) identified by the daytime compartmentalized simulation can function without the FAP guild.
Three in silico consortium analysis methods were developed and applied to a thermophilic phototrophic mat community from Yellowstone National Park as a test case. The results of this systems ecology study demonstrate the applicability of consortium EMA, explicitly mapping the genes-to-function emergent properties associated with the connectivity of metabolic reactions via the exchange of mass and energy. Since the presented techniques are culture-independent, they are compatible with metagenomic approaches and other studies in which culturing or isolation has proven difficult. This study provides a foundation for further work to refine and test the models using co-culture studies; potential inquiries include the measurement of substrate uptake and growth rates, biomass compositions, mRNA and protein expression, and quantification of internal fluxes. Toxicity profiles of the various fermentation products to representative microbes would also be of interest. Representative organisms or enrichment cultures are available for each guild, and the results of such experimentation will be quite useful for model validation and improvement. That work, however, is beyond the scope of the current study, which seeks only to provide methods for the extension of EMA to microbial consortia.
An alternative approach to the mass-balance-based analysis of large networks, flux balance analysis[48, 49], has been applied to genome-scale metabolic networks, which is currently not possible with EMA . Flux balance analysis has been used to investigate central carbon and energy metabolism of a two species consortium . Flux balance analysis could likely be applied to the presented model system and could be used to ask similar questions. The output, however, would be a single flux distribution denoting an optimal use of the network as defined by an objective function, rather than a generating set for all flux distributions allowed by the network (although methods have been established to examine the effects of variations in external constraints on the value of the objective function). As the choice of an appropriate objective function is not always clear prior to extensive experimentation [52, 53], a generating set is useful for exploratory modeling. The utility of a generating set for explanatory modeling has also been recently demonstrated . These benefits justify the development of methods for the application of EMA to multispecies systems alongside large-scale network methods. It is worth considering how the methods developed here will be affected by larger network sizes. This is important because potential applications include more diversified communities and more complex cellular metabolisms. Systems cast as compartmentalized models which prove to be too computationally heavy for current EMA algorithms can be analyzed using the nested approach.
The compartmentalized community analysis method has the advantage of intuitive tractability and separates activity and function by guild, but requires substantially more knowledge of the community than the pooled reactions approach. Tables 1 and 2 illustrate how the energy added to electrons during oxygenic photosynthesis is dissipated with the movement of mass and energy through the consortium. The cyanobacteria exhibit a predicted maximum biomass yield of 0.117 Cmoles for each mole of photons absorbed by the mat. This value is 0.078 for the FAP guild, with the difference providing a rough measure of the cost of transferring reducing equivalents between guilds. The SRB also rely on the by-products of the primary producer guilds for electrons, but are further constrained by an inability to derive energy from light or respiration on O2. The highest SRB biomass yield predicted by the daytime compartmentalized model is 0.016 Cmoles per absorbed mole of photons, slightly more than one tenth of the cyanobacterial yield. The ratio between theoretical optima for photosynthetic and SRB biomass yields is similar to observed ratios between primary producers and the secondary productivity of consumers in macroscale ecological systems [19, 54].
The compartmentalized method also lends itself uniquely to investigation of the robustness of specific consortium interaction types (see Figure 5). Investigating interactions in the model showed that the FAP are central to the metabolic capabilities of the consortium as a whole; FAP involvement was required for the vast majority (99.9%) of elementary modes. Finally, the highest-yielding biomass-producing modes from the compartmentalized approach were combined to show that experimental data supports strategies optimizing productivity by individual guilds more strongly than strategies optimizing productivity for the entire consortium [[9, 16]; Klatt et al., in preparation].
The pooled reactions consortium analysis method modeled community metabolic potential by treating all enzymatic activities and metabolites as residents of the same physical space, without membrane boundaries. The pooled reactions approach represents the coarsest-scale methodology, requires the least a priori information, and is easier to implement than alternative approaches. The pooled approach can often be used when other approaches cannot (due to complexity) or should not (due to lack of detailed data). These advantages are balanced against a tendency to overestimate the metabolic potential. This is unsurprising, as real communities are not super-organisms; individuals are membrane-separated and must contend with the logistics associated with matter and energy transport. The pooled technique is best for initial work on 'poorly' characterized systems. The vast majority of ecosystems are not characterized to a level permitting a compartmentalized analysis, suggesting the pooled approach will be relevant for a long time to come.
Finally, the nested community analysis has properties very similar to the compartmentalized approach, but with the important advantage of easy scalability, achieved by concatenating multiple rounds of EMA analysis. The approach also provides additional ecological insight into the competitive strategies underlying each guild's function. This information is contained within the criteria used to select first-round building blocks. The nested method also easily captures interactions between different guilds as well as between members of the same guild expressing different physiologies. For instance, different cyanobacterial elementary modes can combine, representing exchange of metabolites between cyanobacteria expressing different metabolic activity due to differing positions across spatial gradients of light, temperature, and concentration. This type of exchange appears to be very relevant to actual mat function, as other work suggests the presence of functionally distinct Synechococcus populations that are adapted to particular microenvironments . As with the compartmentalized approach, the daylight case study found that optimal guild strategies were more strongly supported by experimental data than optimal consortium strategies.
Combinations of the pooled and compartmentalized methods can be used to provide information not easily obtainable with any single approach. As an ecologically relevant example, a cost-benefit relationship was quantified for a complete citric acid cycle in the cyanobacterial guild, as opposed to the incomplete cycle indicated by the genomic sequence of relevant isolates. Synechococcus spp. OS-A and OS-B' genomes do not contain the oxaloacetate-producing malate dehydrogenase. The pooled reactions modeling approach identified a cluster of optimal elementary modes using a primarily cyanobacterial metabolism along with the missing citric acid cycle enzyme, which is FAP-derived. The biomass yields in this cluster are a 9.4% improvement relative to purely cyanobacterial metabolic potential, given by the compartmentalized methodology. This is a large difference in yields and laboratory evolution experiments have shown that much smaller differences (0.5% difference in growth rate) are selectable under competitive conditions . These hot spring mat communities are thought to be modern derivatives of ancient prokaryotic communities, and the long history of cyanobacterial and FAP cohabitation suggests the potential for horizontal gene transfer. Evidence for actual horizontal gene transfer events between cyanobacteria and FAP has been published elsewhere . The cyanobacteria in our study have not acquired and passed on the missing Krebs cycle gene, suggesting that the costs of production, maintenance, and regulation for malate dehydrogenase are unlikely to be offset by the benefit of a 9.4% improvement in biomass yield.
Each of the three in silico modeling approaches developed in this study provides a mathematical description of physical constraints on metabolic activity in a consortium. These techniques allow the extension of EMA to ecologically relevant multi-species, biofilm systems. Their contrasting strengths can be combined to arrive at a more holistic description than is possible with any of the methods alone, allowing a broad perspective from which to frame observations and base predictions. These approaches can be adapted to a wide range of microbial communities including both natural and anthropogenic systems. Potential applications include modeling communities involved in wastewater treatment, bioprocess engineering, and environmental remediation, as well as the study of host-pathogen interactions in medicine and symbiotic relationships such as nitrogen-fixing root communities.
The metabolic models presented in this study exist in a control volume consisting of the upper 1 mm of the microbial mat. The steady state control volume considers biomass production but does not include an explicit biomass degradation term. To maintain steady state, biomass would leave the control volume at the same rate it is produced. This would correspond with the physical process of new growth burying old growth. Central carbon and energy metabolism network models for each guild considered to be in the control volume were constructed from literature reviews and annotated genomes of representative organisms including Synechococcus spp. OS-A and OS-B' as well as Roseiflexus sp. RS-1 . GenBank accession numbers are CP000686.1, CP000239.1, and CP000240.1 for the RS-1, OS-A, and OS-B' genomes, respectively. These genomes are highly representative (>90% nucleotide identity to metagenomic reads) of native dominant FAP and cyanobacterial populations in these mats (Klatt et al., in preparation). The metabolic potential of the SRB guild was based on several well-studied organisms (Desulfovibrio vulgaris Hildenborough, Desulfotalea psychrophila, Desulfovibrio desulfuricans G20, Desulfobacterium sp., and Archaeoglobus fulgidus), as well as Thermodesulfovibrio yellowstonii, a thermophilic SRB related to isolates from Mushroom Spring [6, 60–63]. GenBank accession numbers for these genomes, in the stated order, are AE017285.1, CR522870.1, CP000112.1, CP001087.1, AE000782.1, and CP001147.1.
The use of guilds to minimize community complexity is useful for the compartmentalized approach. For the current study, guilds were defined based on four important parameters: energy sources, carbon sources, electron donors and electron acceptors. Each guild utilized a different combination of metabolites to satisfy these parameters. Subdominant guilds were not included in the model (e.g. methanogenic archaea). Here it is assumed that the methanogens would be responsible for degrading phototroph biomass buried within the mat. Since the control volume only considers the active phototrophic region, the decaying biomass would occur outside this control volume removing any possible role played by this subdominant guild.
The metabolic model input consists of the carbon and electron balanced substrates and products for each considered reaction. It should be noted that these organisms are not characterized at a level which would justify genome-scale reproductions. While such models are admirable for well documented organisms, problems with inaccurate, automated genome annotations call into question the benefit of such attempts with any but the best studied microbes . The biomass compositions of these functional guilds have not been experimentally determined due to tight physical coupling of the cells within the microbial mats. The functional guilds reside within a distance of 100 μm. Biomass reactions were written to represent bacterial biosynthetic requirements for a composition of 78% proteins, 16% nucleic acids, and 10% other macromolecules. While the biomass compositions of actual mat constituents were not measured, previous published accounts [e.g. ] as well as unpublished observations from the authors suggest that overall yields are relatively insensitive to the biomass equation as long as the biomass has a biologically relevant degree of reduction (e.g. approximately 4.8 on an N2 basis). The supplemental material contains a degree of reduction analysis of the utilized biomass expression. Its degree of reduction is 4.7. The methods developed here could be applied with alternative biomass synthesis reactions based on future experimental knowledge. Carbon storage consisting of polyglucose or polyhydroxybutyrate was considered separately from biomass production. The metabolic models used in this study are included in additional file 1: Supplement.
Separate models were constructed to represent two distinct phases of a diel cycle, day and night. Differences between the modeled phases were based on the extensive geo- and biochemical data at these sites [5, 11, 12, 55, 65] and included sources and shuttles for carbon and energy as well as oxygen and light availability (see Figure 1 for graphical summary). The day and night metabolisms are connected through the fluxes of reduced carbon and fixed nitrogen. The amount of CO2 fixed during the day determines the mat's overall dark phase reduced carbon budget while the nitrogen fixed during the night using energy stored as polyglucose sets the consortium's fixed nitrogen budget.
Briefly, during the day, the Synechococcus spp. convert solar energy, CO2, and water into oxygen, ATP, and reduced carbon. Concurrently, the anoxygenic FAP derive energy from sunlight and exploit the reduced carbon (glycolate and acetate) produced by the Synechococcus guild. By-products of cyanobacterial activity also serve as carbon and energy sources for SRB. These interactions were used to guide selection of relevant organismal modes for the day nested approach.
At night, the mat was treated as anoxic due to the lack of oxygenic photosynthesis and low oxygen solubility at 60°C. This is a simplification: low levels of oxygen are actually available at the mat surface while most of the mat is anaerobic. The model Synechococcus guild ferments stored polyglucose into a variety of organic compounds including lactate, propionate, acetate, and formate as well as ethanol. In addition to being the primary producer of reduced carbon, Synechococcus is also thought to be the primary producer of fixed nitrogen for the mat . These were the interactions used to guide first-round selection of modes for the night nested method. The nitrogen fixation reactions are considered solely in the night model because of the sensitivity of nitrogenase to oxygen. This could be a simplification of actual mat behavior: mRNA expression evidence and in situ activity assays suggest that a large fraction of nitrogenase synthesis occurs in the morning, when light-derived reducing equivalents are available but before the mat becomes oxic .
The Mushroom and Octopus Spring phototrophic mat communities are found within flowing geothermal springs. Consequently, it was assumed that the mass of each metabolite lost from the control volume due to convective transport was exactly equal to that gained by the control volume through convective transport from upstream microbes. The gases CO2, O2 and H2 were not constrained by this treatment and could vent to the atmosphere.
The nested models require the defining of 'transfer metabolites', which are permitted to serve as metabolic sinks and sources for the individual guild simulations, but not for the community simulation. They were chosen based on knowledge of the springs and compounds commonly transported across bacterial membranes and are shown in Figure 1, associated with arrows between guilds. Individual guild networks were then evaluated for elementary modes with 'transfer metabolites' allowed to accumulate (or deplete). This was accomplished by defining them as external metabolites in CellNetAnalyzer. The resulting modes were sorted in Excel based on product synthesis yields considered relevant to the system. These criteria are listed in additional file 1: Supplement. The modes with high yields, which represent overall guild stoichiometries, were then used as inputs for the second round of EMA (with 'transfer metabolites' now treated as internal metabolites in CellNetAnalyzer). The first round of EMA used enzyme substrate and product stoichiometries to determine how the numerous enzymes can work together as a system, while the second round of EMA used guild stoichiometries to determine how the guilds could work together as a system.
Algorithms for calculating elementary modes from metabolic network topology are described elsewhere . These algorithms are based on a field of mathematics known as convex analysis. Calculation of elementary modes in this work was performed using the software CellNetAnalyzer 9.0 and MATLAB v 7.6 . CellNetAnalyzer is freely available to academics at http://www.mpi-magdeburg.mpg.de/projects/cna/cna.html, but requires MATLAB version 6.1 or higher. Due to computational burden, the models had to be constructed to avoid missing essential enzymatic activity while still permitting a computationally tractable network. Necessary simplifications included rounding, neglecting the accumulation of fermentation products at night, combining macromolecular syntheses into 'lumped' reactions based on draws from the central metabolism, using a proxy for cellular maintenance energy requirements, and streamlining treatment of nitrogen storage. The models presented in additional file 1: Supplement were used to produce all discussed results. A new algorithm for EMA has been recently described that permits the use of a 64 bit Linux based system ; this will likely permit analysis of larger networks than was previously possible. Algorithms for parallel processing have been proposed and will likely reduce the need for simplification, but are not yet publically available .
This study was funded by the NSF IGERT program in Geobiological Systems (DGE 0654336) at Montana State University.
DMW, RG, RPC, and WPI were the authors responsible for conception and design of the experiment. Data was acquired, analyzed, and interpreted by CGK, JEA, KB, NM, RPC, RT, SM, SM, and ZJ. The manuscript was drafted by CGK, JEA, KB, NM, RPC, RT, SM, SM, and ZJ. DMW, RG, RPC, and WPI revised the manuscript critically for intellectual content. All authors read and approved the final manuscript.
| 2019-04-20T22:19:14 |
https://bmcsystbiol.biomedcentral.com/articles/10.1186/1752-0509-3-114
|
0.999943 |
Do you have a book with SAT scores by town for all of Massachusetts?
I didn't think we would have anything in print with scores down to the town level, so I told them I'd search online. They said they had been and couldn't find anything - I told them I'd try anyway, and I'd come get them if I found something (said the librarian, with confidence).
The first place I went was the MA Department of Education website, but a search for SAT scores didn't provide statistics, just news articles about trends.
Next I tried a general web search for SAT scores by town Massachusetts, which did produce a Boston Globe article with scores by schools from 2006. Since this proved such data was available, I thought surely the DOE website must have something.
So I searched again limiting to site:doe.mass.edu (actually, at first I just typed in .gov, but it turns out the DOE website is a .edu - huh), and found the exact same Boston Globe data on the DOE website - plus data from previous years.
It always bugs me when Google's site: limiter search works better than a website's own native search, but at least I found something.
And finally, I searched around CollegeBoard.com to see if they had breakdowns of SAT scores. All I could find there were national percentile tables, but that seems like it might be useful, too.
I went to the study room to tell the kids what I found, and they were pretty happy. Of course, since everything I found was online, the only real way to get it to them was to get the email address of one of the girls in the group and email her the URLs - which seems a little awkward to me, but it worked well enough in this case.
Let me know if you guys need anything else.
They stayed in the room for an hour or so more, and stopped by the desk to thank me on the way out. How nice.
Even though I use Google for web searching most of the time, I do use other search engines, and I wonder how the results compare.
With the launch of Microsoft's new Bing search engine, a Microsoft employee must have been wondering the same thing - so he created a neat Blind Search tool (and states this is not a Microsoft project).
Type in a search term, and Blind Search shows you the results from Google, Yahoo and Bing - but without telling you which engine produced each list. So without brand bias, you decide which results list includes the most relevant websites.
And the best part is the reveal, when you "vote" and see which search engine the results came from.
I played a bit, and surprisingly, Google didn't always provide the most relevant results. As the creator states, this seems most useful as an observational curiosity, but it certainly is fun and interesting (or, it gives people a way to find pron three times faster).
When I use Google to find information, I often use the "site" limiter to improve the returns.
For instance, when looking for information on the new economic stimulus tax rebate thing, a search for "tax rebate site:irs.gov" gives much more direct information than does just searching for "tax rebate." Which is great if you know the domain to which you'd like to limit your search, but yesterday, I didn't.
Someone was looking for information on the James Madison dollar coin, and the U.S. Mint website seemed the most logical place to look for it. However, I didn't know the Mint's domain name. So before my usual site-specific search, I first searched for "us mint" to get the domain, and then I was going to run a second search limited to that domain.
But Google is one step ahead of me (I don't know if this is a new feature or if I just never noticed it before): my search for U.S. Mint returned the Mint's website as the first result, and the listing included a site search built right in to the search result (see picture).
Neat. And it saves me a step. Searching there for "james madison dollar" gave exactly what the patron was looking for as the first result.
I'm generally skeptical of Google as a company for hording private data, but they do have smart people working there.
| 2019-04-19T16:27:46 |
http://www.swissarmylibrarian.net/tag/results/
|
0.999962 |
RUOK Day is on Thursday September 14th.
become a fixture in both calendars and minds across Australia and beyond. September 14th Even though Gavin is no longer alive, his proud legacy continues as ‘RUOK Day’ approaches once again. concern for people with suicidal thoughts calls for courage to ask the question; “RUOK?” and the skills to handle the answer.
The simple question continually delivers many conversations. The RUOK teams travel to public venues like shopping malls across Australia with a simple but powerful resource; how to ask the question and then have the conversation.
The RUOK official site makes the important point that, ibn order to be well placed to support someone else, you must be OK yourself. Self-care then, is the starting point relating to any initiative around concern for people who may be struggling with life. And, just because you may not be the best person to ask someone you’re concerned about if they’re OK, perhaps there’s someone else who might ask the important question and even have a vital conversation.
Not being ready or able to ask the question; RUOK, can simply mean you don’t have the time to spend on a conversation that could be very important for someone’s wellness.
What if the answer is NO!
The question as the RYOK site suggests is the beginning. Most questions generate answers. What if that answer to this one is “no”? It is important then, to know what to do (and what not to) if someone answers; “no”, they are not ok. Maybe it is not the best time for them to discuss their challenges right now, or privacy might be an issue, or the timing is not right somehow. These are all important considerations for any important conversation. Especially this one.
If asking such a question as; RUOK? is uncomfortable for you, consider rehearsing until you get easier with the implications of the possible response. Ask open question to encourage them to speak more and you also show the specific things that caused you to be concerned in the first place. For example, “I notice you didn’t seem yourself at the match yesterday, and I’m just wondering how you’re doing?
Even if you’ve prepared and said your piece, asked your questions, started the conversation, this does not always mean they’re ready to open up and talk. This does not mean you’ve failed or are in any way responsible for their behaviour. Leaving the door open for talk later on is still worthwhile. Checking if they’d like to talk to someone else, reminding them that you still care, are still worthwhile.
Listening is an art. Listening well and paying good attention can be powerful resources in attempts to support someone in pain. Avoid finishes their sentences, rushing them with head-nods or other gestures and don’t judge. Silence itself is golden sometimes. Giving ‘space’ to someone so they can process their understanding of their experience before they speak about how they feel can be powerfully supportive. Encouraging elaboration on something they’ve said can also communicate that you’ve been listening well and paying respectful attention.
| 2019-04-23T22:59:04 |
http://bondihypnotherapyclinic.com.au/alright-or-ruok/
|
0.999999 |
A Florida man buried his friend's body in the backyard of his house and inadvertently told the police about it when they came to visit him on an unrelated issue.
Homeowner Greg Palermo, 50, from Orange County, says he & # 39; has made a bad decision & # 39; when he buried the man six meters below him where his patio was.
The police at the Sheriff & # 39; s Office in Orange County finally found the body on Sunday in the courtyard of the house in the Doctor Phillips district of Orlando after Palermo told them the exact location.
While talking to a local television station, Palermo explained that his friend, who lived in the house from time to time, had died days earlier from an overdose of drugs.
Palermo panicked that he could blame the death of his friends and decided to bury him in the backyard himself.
Palermo claimed that his concerns about alerting the police to the death of the friend at home stemmed from the fact that he had previously been arrested for drug charges.
But in the end he accidentally told them, after he had asked the police how they knew they should look for a corpse in his backyard.
In reality, the police had no idea there was a corpse a meter away, but by that time Palermo had given away the game with his innocent questions.
The police had actually come to the house in an unrelated case and went looking for possible drug paraphernalia.
& # 39; How did law enforcement know about it? I told them they came here because they thought I was a manufacturer of meth and I don't, & Palermo said.
No charges have been lodged against Palermo in the light of the new information.
The Orange County Sheriff & # 39; s Office also did not comment on the matter.
| 2019-04-23T04:57:25 |
https://whatsnew2day.com/a-bizarre-moment-tells-a-man-from-florida-to-a-tv-team-where-he-buried-the-body-of-his-friend-in-the-backyard/
|
0.999998 |
Why is the UK seeing a record fall in workers’ wages?
Real wages are falling at a near-record rate. Wednesday's figures show that they were 6% lower in April than they were in April 2008.
This is the biggest five-year drop in real wages since 1921-26, and the second-largest fall since records began in 1855.
This cannot be blamed simply on the recession. As the IFS has pointed out (pdf), real wages rose during the recessions of the early 80s and 90s. Something, then, has changed since then. But what?
Here's a theory. Back in the 70s and 80s, bosses could often not efficiently monitor their workers. To keep pilfering and skiving within tolerable limits they therefore had to pay better than market-clearing wages, to buy goodwill. The upshot was that wages rose even during downturns, because bosses feared that real wage cuts would create discontent and thus increase thieving, insubordination and malingering.
This led to a huge literature in economics on efficiency wages, gift exchange (pdf) and insider-outsiders (pdf), which tried to explain high and sticky real wages.
However, as Frederick Guy and Peter Skott have shown, socio-technical change since the 80s such as CCTV, containerization and computerized stock control has made it easier for bosses to monitor workers. Direct oversight means they don't need to worry about buying workers' goodwill. They are instead using the Charles Colson strategy: "When you've got 'em by the balls, their hearts and minds will follow."
Years ago, firms wanted smaller but motivated workforces. Now they can control workers directly, they don't need to worry so much about motivation* and so are content with larger but grumpy workers.
1. Talk of "wage rage" misses an important point. At the point of production – to use a quaint Marxian phrase – there is little meaningful rage, because workers can do little to fight falling real wages. (This poses the danger that such rage will find perhaps misdirected political expression, such as in antipathy towards immigrants).
2. Issues of industrial organization – how firms are organized – have important macroeconomic effects. Macroeconomics cannot be easily studied separately from ind. org. Economists need to look inside the "black box" of industrial structure.
3. You cannot understand economics without understanding power. The fact is that bosses' power has risen and (many) workers' power has declined. In this sense, the rising incomes of the 1% and the fall in real wages for the average worker are two manifestations of the same process.
* except, of course for top-level managers who cannot be directly monitored – hence their rising incomes.
this theory of does tend to miss out that sometime in the 80s corporate profits went from an average of 8 – 10% to 25 – 30%. This wasn’t just shareholder greed, although that certainly helped but was brought about mostly by the twin threat from banks who used catch-all head office determined formulae about profit criteria which prevented them lending to “under-performing” enterprises. And companies which couldn’t get loans were at the mercy of the new financier game of hedge fund leveraged buyouts.
All of whom protested that they were simply doing this, almost altruistically, to “eliminate inefficiency” and “liberate value”. Or closing down plant and shedding workforce.
In Germany c2000, wages were too high in relation to the production coming from those being paid. Productivity was too low. So, the German government engineered a relative fall in the workers’ wages. As productivity continued to rise wages did not, thus reducing the labour cost of production. And after a decade of doing this German industry is now extremely competitive.
British industry faces much the same problem. Wages are simply too high compared to the output from employing people. Thus the wages need to fall relative to output. Which is what seems to be happening. Hurrah!
…and as wages fall – demand falls as the worker can no longer afford the goods and services he’s producing.
Businesses go bust – more jobs are lost and the world declines into a deep depression.
If only employers thought about the consequences of falling wages rather than just blindly accuming they mean ‘more profit’ for the greedy shits.
That’s true, in broad way. It is also true that lower wages reduce the ability to save or get a mortgage, which changes housing costs. Many goods are pitched at the market on the basis of “what consumers can afford/will pay” rather than cost of manufacture and production; reduced wages lower the prices of such goods. When working people have less money, they spend it on different things. Spending patterns change, often in counter intuitive ways; UK consumers currently buy more good wine (absolute and as a percentage) than 15 years ago.
Demand falls but businesses and consumers also change. Whether/how this results in fewer jobs or reduction in quality of life is more complicated.
But that’s simply not what happened in Germany.
As wages fall, many prices fall, too – all other things being equal. Moreover, as wages fall, employment increases, and fewer people are unemployed than would otherwise be the case. So, broadly, demand can be maintained during the adjustment to higher productivity.
Would really like you to list some examples of ‘Hedge fund leveraged buy outs’. Profits at 25 -30% of turnover? I doubt that lasted for very long.
Obviously employers should immediately triple the wages of all of their workers and everyone will be better off.
You are missing the obvious decline of unions in the period between 1970s and now. Individual has replaced collective bargaining which puts low skilled workers in particular, at a disadvantage. This has probably been exacerbated by the apparent willingness of women, who constitute a larger proportion of the workforce than in the past, to accept lower wages than men for the same work.
The other, connected, change is to service industries whose organization (outside financial services) dispenses with tiers of skilled workers and pays a single rate for unskilled work, and a single rate is more easily eroded than multiple rates.
There is also, I think, more pressure from shareholders meaning that costs are screwed to the floor, and in service industries wages are usually the most significant cost.
@ 2 Its not that productivity is too low it’s that prices are too high. If you want to drop wages then prices have to fall, otherwise the workforce/consumers can’t afford to buy anything. That is what is happening at the moment, 2007 real prices on 2013 real wages, so demand falls. The response should be prices fall but it isn’t, because that would harm the interests of management.
This is the real problem. UK industry has always and still does suffer from chronic over management. Too many managers as a ratio of workers, they are not on average very good and they are on average overpaid. If you want BRIC productivity you have to have the flip side which is BRIC company structures and BRIC remuneration for the board.
If we had BRIC productivity we would be massively poorer because BRIC productivity is hugely below UK productivity, which is kinda why they are poor. Productivity has very little to do with how hard someone works, but is related to the output they produce. Someone digging ditches with a shovel will be working hard. An earth moving operator will not be working so hard but will be more productive.
I think there needs to be more disaggregating done rather than just look at the issue of wages as a homogenous whole. When wages are disaggregated it usually is the case that it is manual male wages that have been devalued. The decline of unions may partially provide some explanation but it is lazy to think of that as a theory of everything. The same phenomenon happened all over the advanced industrial world no matter what happened with union membership. Moreover, most of the type of labour that has been devalued never had high union membership anyway. Marxists always see things in terms of power relations and it probably is the case that labour that can easily be automated does not hold much power.
Indeed, but initially at least automation tends to replace low skilled work. This is what preceded the feudal/industrial transition. Subsequently, ICT has done the same, or should have, to divisional company structures. The company of the future will be a board, some contract technocrats, and an outsourced workforce. It’s already practically happened in a lot of the service industries. What is the point of all those middle managers synthesising and analysing data when a computer can do it cheaper and quicker. The pressure for this is coming from above (i.e shareholders), so Marxist or not, you cannot ignore power relations.
How come this in the news?
Could it be that immigrant males are better educated and more productive than unemployed indigenous males?
Obviously, printing money doesn’t effect anything doesn’t cause inflation. Inflation is only measured when the government participates in the economy because it effects prices for business’ hey.
How much concrete can the bankers pour in this country?
The Alantic a bit to left wing for people believe printing money is the best way to run a country.
As you can see Sth Korea has the highest wage increases and good jobs. Well Done David Cameron!!!!!!
Yeah Milton Friedman can print money in England to buy Detroit. A few articles in relation to hedge funds and why they are a fraud.
I’m sure many of them are quite aware of that, but there’s not a lot you can do if your competitors are cutting wages in order to cut prices. In most markets, if you don’t keep up with the competition on price you’re going to get destroyed.
It’s a form of market failure. Lower wages are advantageous for almost every single business in the market, but where the earners are also the consumers, ultimately damaging for the market as a whole.
Some raised a comparison with Germany – and it is interesting to note the very different fortunes of the UK and Germany in recent years, because it emphasizes how badly the UK worker is doing.
Germany has not seen a similar recent dip – wages are down by about 3% on 2000, but that is for longer term reasons that saw negotiated pay reductions between unions and manufacturers under Schroder. This is particularly surprising as Germany appears even now to be running close to structurally full employment (about 5% unemployment is likely to be as close to effective 0 unemployment as a modern economy gets) – so its average wage is not boosted at all by dumping lots of low wage workers onto the unemployment column.
So why is Germany faring so much better? Partly this is likely to be far more structural than the UK is ever able to act on. Germany has a far more diverse economy than the UK’s rather service-dependent economy. Services often need generic skills so hiring and firing is easy, as is pushing down wages – while Germany has more specialist skilled workers, who are by nature of the training and education involved, harder to replace and more valuable in terms of productivity.
To fix that the UK would have to move to a more diverse economic model – and moderiise our education system dramatically (for both kids and adults). Since we have been cutting adult education for some time and given that our education secretary is trying to un-modernise our education system for kids – it seems unlikely we are set to follow the German model no matter how successful it is.
As such we should expect our nation to keep getting poorer.
| 2019-04-22T13:24:24 |
http://liberalconspiracy.org/2013/06/14/why-is-the-uk-seeing-a-record-fall-in-workers-wages/
|
0.999996 |
Information on Mary Shaw who married Reverend Joseph Burson?
What was the last name of Leah who married Joseph Jefferson Burson?
1. GEORGE1 BURSON was born Abt. 1624 in Scarborough, Yorks, Eng..
i. JOSEPH2 BURSON, b. 1650; d. 1700, founder of Quakertown, Pa; m. GEMIMA STROUD.
2. ii. GEORGE BURSON, b. Abt. 1662, Scarborough, England, Quaker; d. November 26, 1715, Gilbert's Manor, Montgomery, Pennsylvania.
2. GEORGE2 BURSON (GEORGE1) was born Abt. 1662 in Scarborough, England, Quaker, and died November 26, 1715 in Gilbert's Manor, Montgomery, Pennsylvania. He married HANNAH GOULD.
i. JAMES3 BURSON, b. September 29, 1688, abington, Montgomery, Pennsylvania.
3. ii. JOSEPH BURSON, b. December 25, 1689, Gilbert's Manor, Montgomery, Pennsylvania; d. Aft. October 19, 1769, Loudon, Virginia.
iii. MARY BURSON, b. 1692, abington, Montgomery, Pennsylvania.
iv. ANN BURSON, b. 1694, abington, Montgomery, Pennsylvania.
v. WILLIAM BURSON, b. 1696, abington, Montgomery, Pennsylvania.
vi. GEORGE BURSON, b. 1698, Abington, Montgomery, Pennsylvania; d. 1786, Loudon, Virginia; m. SARAH COX, September 1731, Gilbert's Manor, Montgomery, Pennsylvania.
3. JOSEPH3 BURSON (GEORGE2, GEORGE1) was born December 25, 1689 in Gilbert's Manor, Montgomery, Pennsylvania, and died Aft. October 19, 1769 in Loudon, Virginia. He married MARY RACHEL POTTS October 15, 1719 in Gilbert's Manor, Philadelphia Pennsylvania, daughter of JONAS POTTS and MARY THOMAS.
4. i. ANN4 BURSON, b. February 09, 1720/21, Gilbert's Manor, Montgomery, Pennsylvania.
ii. JAMES BURSON, b. November 21, 1722, Gilbert's Manor, Philadelphia Pennsylvania; d. 1792; m. (1) SARAH WORTH; m. (2) SARAH PRICE.
5. iii. JOSEPH BURSON , REV., b. September 1730, Gilbert's Manor, Philadelphia Pennsylvania, Bucks Co.; d. Aft. October 28, 1801, Warren County, Georgia.
iv. RACHEL BURSON, b. 1732, Gilbert's Manor, Montgomery, Pennsylvania.
v. BENJAMIN BURSON, b. 1733, Gilbert's Manor, Philadelphia Pennsylvania; d. 1769, Will in Loudon Co., Va.; m. ANNE DICKEY.
vi. BURSON BURSON, b. 1735.
vii. DEBORAH BURSON, b. 1737.
viii. SARAH BURSON, b. 1739.
4. ANN4 BURSON (JOSEPH3, GEORGE2, GEORGE1) was born February 09, 1720/21 in Gilbert's Manor, Montgomery, Pennsylvania. She married ISAAC GRIFFITH November 1744 in Buckingham MM, Bucks, Pennsylvania.
ii. ABRAHAM GRIFFITH, b. January 05, 1745/46, Richland, Phil. PA.
iii. RACHEL GRIFFITH, b. November 23, 1747, Richland, Phil. PA; m. JOSEPH SHAW.
iv. HANNAH GRIFFITH, b. September 06, 1749; m. MOSES DILLON.
v. JOSEPH GRIFFITH, b. October 22, 1750, Richland, Bucks Pennsylvania.
vi. MARY GRIFFITH, b. July 28, 1752, Richland, town, Bucks Pennsylvania; m. JOHN NAYLOR.
vii. ANN GRIFFITH, b. February 01, 1754, Richland, Bucks Pennsylvania.
viii. SARAH GRIFFITH, b. November 15, 1756, Richland, Bucks Pennsylvania.
ix. MARTHA GRIFFITH, b. February 23, 1760, Richland, Bucks Pennsylvania.
x. ISAAC GRIFFITH, b. June 07, 1764.
xi. JOHN GRIFFITH, b. May 24, 1766.
5. JOSEPH4 BURSON , REV. (JOSEPH3, GEORGE2, GEORGE1) was born September 1730 in Gilbert's Manor, Philadelphia Pennsylvania, Bucks Co., and died Aft. October 28, 1801 in Warren County, Georgia. He married MARY SHAW 1750 in Bucks County, Pa. Pennsylvania.
i. JONATHAN5 BURSON, b. Abt. 1750, Fairfax Co., Virginia Colony; d. April 14, 1801, Warren Co., Georgia, USA; m. (1) JANE WHINNERY; m. (2) NANCY GREEN, 1778, Old 96th District, SC, USA.
6. ii. ISAAC BURSON , REV. WAR, b. 1752, Louden, Virginia; d. 1829, Jackson, Georgia.
iii. PEGGY BURSON, b. 1753, 96th, S.C..
iv. POLLY BURSON, b. 1755, 96th, S.C..
7. v. LYDIA BURSON, b. 1756, Loudoun Co., Virginia Colony; d. Abt. 1841, Castle, Cass Co., Indiana, USA.
vi. PHOEBE BURSON, b. 1759, 96th, S.C..
vii. PEGGY BURSON, b. 1764, Loudoun Co., Virginia Colony.
viii. POLLY BURSON, b. 1767, South Carolina Colony.
ix. RACHEL BURSON, b. March 08, 1768, Union County, S. C. .; m. JOHN NEAL, May 25, 1790, Georgia, USA.
8. x. JOSEPH JEFFERSON BURSON, b. 1771, Old 96th Dst., South Carolina; d. Bef. July 14, 1851, Wilcox County, Alabama.
xi. JESSE BURSON, b. 1774, 96th, S.C..
xii. ENOCH BURSON, b. 1775, 96th, S.C..
xiii. ALICE BURSON, b. 1777, 96th, S.C..
6. ISAAC5 BURSON , REV. WAR (JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1752 in Louden, Virginia, and died 1829 in Jackson, Georgia. He married SARAH SHAW 1774 in Union Co., South Carolina Colony.
i. DAVID6 BURSON, b. Walton, Georgia; d. Bef. 1830; m. ELIZABETH ROSS?.
9. iii. JOSEPH BURSON, b. Abt. 1776.
iv. ISAAC BURSON, b. 1788, South Carolina; m. SARAH.
v. SARAH (SALLY) BURSON, b. Abt. 1790, Jackson, Georgia; m. WOOD LEROY HINTON, September 20, 1811, Jackson, Georgia.
vi. DELILAH BURSON, b. Abt. 1790, Jackson, Georgia; m. ELIJAH SHAW, January 17, 1809.
vii. ELISHA BURSON, b. 1790.
viii. SILVANEY BURSON, b. 1799.
7. LYDIA5 BURSON (JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1756 in Loudoun Co., Virginia Colony, and died Abt. 1841 in Castle, Cass Co., Indiana, USA. She married WILLIAM YORK 1781 in North Carolina, USA.
i. AARON6 YORK, b. 1782; d. 1820, Ohio, USA.
ii. JOSEPH YORK (2), b. 1783, Wilkes Co., Georgia, USA; d. 1783, Wilkes Co., Georgia, USA.
iii. MARY YORK, b. 1783, Wilkes Co., Georgia, USA; d. Putnam Co., Indiana, USA; m. THOMAS J. MOORE, Georgia, USA.
10. iv. NEWBERRY YORK, b. September 06, 1784, Wilkes Co., Georgia, USA; d. March 31, 1866, York Twp., Darke Co., Ohio, USA.
v. WILLIAM YORK (3), b. 1786, Wilkes Co., Georgia, USA; d. 1799, Warren Co., Georgia, USA.
vi. JEPTHA YORK (2), b. April 14, 1788, Wrightsboro, McDuffie Co., Georgia, USA; d. February 09, 1846, White Co., Indiana, USA; m. SUSANNA SLADE, March 23, 1808, Preble Co., Ohio, USA.
vii. MAHALA YORK, b. March 25, 1793, Columbia Co., Georgia, USA; d. 1828, Preble Co., Ohio, USA; m. JOHN R. WILLIAMS, February 22, 1810, Preble Co., Ohio, USA.
viii. CLEMENTINE YORK, b. August 04, 1795, Warren Co., Georgia, USA; d. January 12, 1887, Travis Co., Texas, USA; m. JAMES BRATTON ELLIOTT , JR., September 27, 1818, Preble Co., Ohio, USA.
ix. DIADAMA YORK (2), b. 1797, Warren Co., Georgia, USA; d. September 17, 1858, Cass Co., Indiana, USA; m. JOHN WINEGARDNER, October 21, 1814, Preble Co., Ohio, USA.
x. CLARISSA YORK, b. 1802, Warren Co., Georgia, USA; m. JOHN KING MATTOX, August 05, 1830, Preble Co., Ohio, USA.
8. JOSEPH JEFFERSON5 BURSON (JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1771 in Old 96th Dst., South Carolina, and died Bef. July 14, 1851 in Wilcox County, Alabama. He married (1) LEAH February 06, 1792 in Georgia, USA. He married (2) WINNIE A. July 15, 1843.
i. KATIE6 BURSON, b. May 16, 1793; d. December 1810.
11. ii. AMOS BURSON, b. August 21, 1794.
iii. ENOCH BURSON, b. July 18, 1796.
iv. SOLOMON BURSON, b. January 23, 1799.
v. MARY BURSON, b. January 23, 1800.
vi. JESSE BURSON, b. October 07, 1802.
12. vii. JOSEPH (JACKSON) BURSON, b. February 11, 1804, Georgia; d. December 10, 1873, buried in Burson cemetery in Pine Barren near cabin.
viii. WILLIAM JEFFERSON BURSON, b. March 20, 1806; d. August 14, 1816, Tennessee.
ix. ISAAC BARTLETT BURSON, b. February 15, 1808; d. January 30, 1911, Pineville, Alabama.
x. JONATHAN FRANKLIN BURSON, b. April 14, 1810; d. Wilcox County, Enon Cemetery, Alabama; m. MARGARET SMITH.
xi. RICHARD AUGUSTUS BURSON, b. April 06, 1813; d. July 24, 1886, Pineville, Alabama; m. HARRIETTA JANE DREW, February 13, 1839.
xii. ALMIRA BURSON, b. July 20, 1815.
9. JOSEPH6 BURSON (ISAAC5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born Abt. 1776.
i. MARY A.7 BURSON, b. August 08, 1807, Georgia; m. BENJAMIN KING, 1828, Jackson, Georgia.
ii. GEORGE W. BURSON, b. 1814; d. Aft. 1884, Carroll, Georgia.
13. iii. JAMES C. BURSON, b. June 17, 1816; d. July 05, 1878, Woods family cemetery near J.T. Cotton place on Garrett Ferry Road.
10. NEWBERRY6 YORK (LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born September 06, 1784 in Wilkes Co., Georgia, USA, and died March 31, 1866 in York Twp., Darke Co., Ohio, USA. He married NANCY A. SLADE January 05, 1804 in Wrightsboro, Johnson Co., Georgia, USA.
14. i. JOSEPH7 YORK, b. October 12, 1804, Wilkes Co., Georgia, USA; d. November 22, 1862, Mercer Co., Ohio, USA.
15. ii. JEREMIAH YORK, b. April 07, 1809, Preble Co., Ohio, USA; d. January 27, 1862, Wayne Twp., Darke Co., Ohio, USA.
16. iii. NICHOLAS SLADE (ESQ.) YORK, b. July 09, 1812, Montgomery Co., Ohio, USA; d. July 01, 1901, York Twp., Darke Co., Ohio, USA.
17. iv. DIADAMA YORK, b. November 12, 1815, Preble Co., Ohio, USA; d. 1885, Mercer Co., Ohio, USA.
18. v. WILLIAM YORK (5), b. November 17, 1816, Wilkes Co., Georgia, USA; d. December 21, 1877, Darke Co., Ohio, USA.
vi. JEPTHA YORK, b. September 11, 1818, Preble Co., Ohio, USA; d. August 13, 1843, Darke Co., Ohio, USA; m. ELIZA JANE WILLIS, December 10, 1841, Darke Co., Ohio, USA.
vii. LEWIS YORK, b. November 20, 1822, Darke Co., Ohio, USA.
19. viii. REBECCA YORK, b. April 12, 1824, Darke Co., Ohio, USA; d. September 01, 1907, Darke Co., Ohio, USA.
ix. NEWBERRY E. YORK, b. January 24, 1848, Darke Co., Ohio, USA; d. January 22, 1876, Darke Co., Ohio, USA; m. MARY GILBERT, October 20, 1856, Darke Co., Ohio, USA.
11. AMOS6 BURSON (JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born August 21, 1794. He married (1) NARCISSUS NEAL. He married (2) SARAH NEAL March 22, 1822.
12. JOSEPH (JACKSON)6 BURSON (JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 11, 1804 in Georgia, and died December 10, 1873 in buried in Burson cemetery in Pine Barren near cabin. He married ELIZABETH DUNN February 14, 1829, daughter of BENJAMIN DUNN and SUSANNAH.
21. i. MARTHA ANN7 BURSON, b. July 02, 1830; d. January 22, 1868, buried in Burson cemetery in Pine Barren near cabin.
22. ii. ELKANAH BURSON , CIVIL WAR, b. January 23, 1832, Snow Hill, Wilcox County, Alabama; d. February 11, 1915, Furman, Alabama, Wilcox County, Palmer Cem..
iii. ALFRED BURSON, b. October 07, 1836; d. October 15, 1915, buried in Burson cemetery in Pine Barren near cabin.
iv. ARVILLA BURSON, b. June 01, 1840; d. June 01, 1852.
13. JAMES C.7 BURSON (JOSEPH6, ISAAC5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born June 17, 1816, and died July 05, 1878 in Woods family cemetery near J.T. Cotton place on Garrett Ferry Road. He married LUCINDY POLK?.
23. i. JAMES POLK8 BURSON.
14. JOSEPH7 YORK (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born October 12, 1804 in Wilkes Co., Georgia, USA, and died November 22, 1862 in Mercer Co., Ohio, USA. He married MARY (POLLY) BAKER June 30, 1825 in Darke Co., Ohio, USA.
i. ELINOR8 YORK, b. June 24, 1826, Darke Co., Ohio, USA.
24. ii. NANCY YORK, b. December 25, 1827, Darke Co., Ohio, USA.
iii. NEWBERRY YORK (2), b. October 16, 1829, Darke Co., Ohio, USA.
25. iv. DELILA YORK, b. August 23, 1831, Darke Co., Ohio, USA.
v. DIADAMIA YORK, b. January 27, 1833; m. MICHAEL ADAMS, January 22, 1863.
26. vi. LOUISA YORK, b. October 19, 1834; d. April 23, 1878.
vii. ADAM YORK, b. November 19, 1836.
viii. LEWIS YORK (2), b. November 19, 1836; d. January 26, 1874.
27. ix. ELMIRA YORK, b. 1840.
28. x. REBECCA YORK (2), b. 1842.
29. xi. EVAN YORK, b. 1848; d. 1937.
15. JEREMIAH7 YORK (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 07, 1809 in Preble Co., Ohio, USA1,2, and died January 27, 1862 in Wayne Twp., Darke Co., Ohio, USA. He married (1) NANCY WOODS (3) June 28, 1830 in Darke Co., Ohio, USA. He married (2) DIADEMIA MENDENHALL October 10, 1846 in Darke Co., Ohio, USA.
30. i. MARY8 YORK (2), b. December 22, 1830, Darke Co., Ohio, USA; d. August 09, 1894, Darke Co., Ohio, USA.
ii. LUELLA YORK, b. November 07, 1832.
31. iii. LYDIA YORK, b. November 07, 1832, Darke Co., Ohio, USA.
32. iv. CATHERINE ANN YORK, b. March 31, 1835, Darke Co., Ohio, USA; d. December 28, 1855, Darke Co., Ohio, USA.
33. v. JOHN G. YORK, b. February 03, 1836, Darke Co., Ohio, USA; d. November 08, 1908, Darke Co., Ohio, USA.
34. vi. ELIZA JANE YORK, b. May 08, 1837, Darke Co., Ohio, USA; d. May 28, 1858, Darke Co., Ohio, USA.
35. vii. CAROLINE YORK, b. May 11, 1838, Darke Co., Ohio, USA; d. March 29, 1916, Darke Co., Ohio, USA.
viii. UNKNOWN YORK (23), b. February 1841, Darke Co., Ohio, USA; d. November 1845, Darke Co., Ohio, USA.
ix. SEREPTA YORK (40), b. March 07, 1844, Darke Co., Ohio, USA; d. August 02, 1872, Darke Co., Ohio, USA; m. (1) MR. REED, Aft. 1860, Darke Co., Ohio, USA; m. (2) CHARLES E. MENDENHALL, September 26, 1866, Darke Co., Ohio, USA.
x. NANCY YORK (9), b. November 09, 1845, Darke Co., Ohio, USA; d. November 09, 1845, Darke Co., Ohio, USA.
xi. RICHARD8 YORK (2), b. June 12, 1847, Wayne Twp., Darke Co., Ohio, USA3,4; d. June 20, 1911, Eastlake, Tennessee, USA; m. OLIVE M. RICHARDSON, June 27, 1868, Darke Co., Ohio, USA.
xii. NEWBERRY YORK (3), b. January 23, 1849, Wayne Twp., Darke Co., Ohio, USA5,6; d. August 20, 1856, Wayne Twp., Darke Co., Ohio, USA.
36. xiii. WILLIAM YORK (4), b. February 02, 1851, Wayne Twp., Darke Co., Ohio, USA; d. December 13, 1940, Muncie, Delaware Co., Indiana, USA.
xiv. SARAH E. YORK, b. March 19, 1853, Wayne Twp., Darke Co., Ohio, USA7,8; d. December 26, 1861, Wayne Twp., Darke Co., Ohio, USA.
xv. NANCY YORK (8), b. December 28, 1854, Wayne Twp., Darke Co., Ohio, USA9,10; d. Abt. 1863, Wayne Twp., Darke Co., Ohio, USA.
37. xvi. RACHAEL YORK, b. May 27, 1856, Wayne Twp., Darke Co., Ohio, USA; d. March 02, 1932, Versailles, Wayne Twp., Darke Co., Ohio, USA.
xvii. PRISCILLA J. YORK, b. April 28, 1858, Wayne Twp., Darke Co., Ohio, USA11,12; d. September 20, 1859, Wayne Twp., Darke Co., Ohio, USA.
xviii. JEREMIAH YORK (2), b. March 11, 1860, Wayne Twp., Darke Co., Ohio, USA13,14; d. July 29, 1860, Wayne Twp., Darke Co., Ohio, USA.
16. NICHOLAS SLADE (ESQ.)7 YORK (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 09, 1812 in Montgomery Co., Ohio, USA, and died July 01, 1901 in York Twp., Darke Co., Ohio, USA. He married JEMIMA WARD March 13, 1834 in Darke Co., Ohio, USA.
38. i. JOSEPH SLADE8 YORK, b. July 10, 1835, Darke Co., Ohio, USA; d. July 22, 1929, Darke Co., Ohio, USA.
39. ii. SARAH JANE YORK, b. November 06, 1837, Darke Co., Ohio, USA; d. November 17, 1911, Ohio, USA.
40. iii. ELIJAH YORK, b. July 26, 1840, Darke Co., Ohio, USA; d. December 01, 1868, Darke Co., Ohio, USA.
41. iv. IRVIN YORK, b. September 26, 1842, Darke Co., Ohio, USA; d. March 28, 1910, Darke Co., Ohio, USA.
42. v. DAVID YORK, b. October 18, 1844, Darke Co., Ohio, USA; d. October 05, 1925, Darke Co., Ohio, USA.
vi. JOHN T. YORK, b. May 29, 1847, Darke Co., Ohio, USA; d. September 03, 1848, Darke Co., Ohio, USA.
43. vii. WILLIAM CALVIN YORK, b. August 15, 1850, Darke Co., Ohio, USA; d. January 20, 1935, Shelbyville, Shelby Co., Indiana, USA.
viii. NANCY ANN YORK, b. March 17, 1854, Darke Co., Ohio, USA; d. June 06, 1854, Darke Co., Ohio, USA.
44. ix. NEWBERRY FRANKLIN YORK, b. January 10, 1857, Darke Co., Ohio, USA; d. October 21, 1921, Ansonia, Brown Twp., Darke Co., Ohio, USA.
17. DIADAMA7 YORK (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 12, 1815 in Preble Co., Ohio, USA, and died 1885 in Mercer Co., Ohio, USA. She married JAMES T. MILLER September 09, 1833 in Darke Co., Ohio, USA.
i. HIRAM8 MILLER, b. March 14, 1837; d. July 08, 1868; m. SARAH SMITH, October 21, 1858.
ii. SUSANNAH MILLER, b. January 1838; m. MARTIN J. MARKER, June 18, 1862.
45. iii. NEWBERRY F. MILLER, b. May 20, 1840; d. September 21, 1921.
iv. MARTHA JANE MILLER, b. Abt. 1842; d. August 12, 1897; m. MARTIN LEHMAN, May 04, 1871.
v. JEREMIAH MILLER, b. Abt. 1844; m. MARY BOOMERSHIRE, December 31, 1871.
vi. CATHERINE MILLER, b. Abt. 1846; m. SAMUEL LENICH, February 18, 1869.
vii. ELIZABETH MILLER, b. Abt. 1848; d. 1912; m. WILLIAM J. ROSS, September 18, 1854.
viii. NANCY MILLER, b. Abt. 1850; m. WILLIAM NUTTER OR HUNTER, April 08, 1875.
18. WILLIAM7 YORK (5) (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 17, 1816 in Wilkes Co., Georgia, USA, and died December 21, 1877 in Darke Co., Ohio, USA. He married CATHERINE WHITMAN September 24, 1829 in Darke Co., Ohio, USA.
46. i. ROSETTA8 YORK, b. July 12, 1830; d. October 11, 1877.
47. ii. NANCY YORK (2), b. September 21, 1833; d. February 07, 1865.
iii. LUCRETIA YORK, b. 1836; m. MANSFIELD BAYMAN, November 13, 1859.
48. iv. NICHOLAS J. YORK, b. Abt. 1838; d. 1908.
v. SARAH ANN YORK, b. March 31, 1842; d. May 09, 1872.
49. vi. EFFA YORK, b. 1844.
19. REBECCA7 YORK (NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 12, 1824 in Darke Co., Ohio, USA, and died September 01, 1907 in Darke Co., Ohio, USA. She married SAMUEL ARMSTRONG December 01, 1842 in Darke Co., Ohio, USA.
i. JEPTHA8 ARMSTRONG, b. 1843, Darke Co., Ohio, USA; d. 1910.
ii. HANNAH ARMSTRONG, b. 1846, Darke Co., Ohio, USA; d. September 26, 1849, Darke Co., Ohio, USA.
50. iii. JAMES D. ARMSTRONG, b. 1848, Ohio, USA; d. 1927.
iv. NANCY JANE ARMSTRONG, b. 1850, Darke Co., Ohio, USA; d. 1916.
v. JOBE ARMSTRONG, b. February 28, 1852, Darke Co., Ohio, USA; d. December 05, 1920, Darke Co., Ohio, USA.
vi. NEWBERRY ARMSTRONG, b. 1854, Darke Co., Ohio, USA; d. 1934.
vii. SAMUEL ARMSTRONG (3), b. September 16, 1855, York Twp., Darke Co., Ohio, USA; d. March 16, 1876, York Twp., Darke Co., Ohio, USA.
viii. DIADAMA ARMSTRONG, b. September 01, 1857, York Twp., Darke Co., Ohio, USA; d. October 28, 1875, Darke Co., Ohio, USA.
ix. ENOS ARMSTRONG, b. 1859, York Twp., Darke Co., Ohio, USA; d. 1928.
x. CLARA ARMSTRONG, b. 1863, Darke Co., Ohio, USA; d. 1956.
20. MARY7 BURSON (AMOS6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1). She married JAMES JEFFERSON ADAMS.
21. MARTHA ANN7 BURSON (JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 02, 1830, and died January 22, 1868 in buried in Burson cemetery in Pine Barren near cabin. She married JAMES L. HURST December 09, 1860.
51. i. LULA BELL8 HURST, b. September 27, 1861; d. May 31, 1929.
52. ii. WILLIAM JAMES JACKSON HURST, b. December 22, 1863.
53. iii. ELKANAH LAWRENCE EARLY HURST, b. May 09, 1867.
22. ELKANAH7 BURSON , CIVIL WAR (JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born January 23, 1832 in Snow Hill, Wilcox County, Alabama, and died February 11, 1915 in Furman, Alabama, Wilcox County, Palmer Cem.. He married ELEFARE CHRISTIAN BARGE January 23, 1876, daughter of GEORGE BARGE and MARY SLAUGHTER.
i. MARY ELIZABETH8 BURSON, b. April 10, 1880.
54. ii. ELKANAH GEORGE BURSON DR., b. April 07, 1882, Furman, Wilcox County, Alabama; d. April 1970, Furman, Wilcox County, Alabama buried in Palmer Cemetery.
iii. ELLA CLARE BURSON, b. February 03, 1886; d. June 23, 1993, Dothan, Aabama Houston County; m. (1) G.A. UNDERWOOD, May 19, 1919; m. (2) JAMES LEWIS MOULDER, June 09, 1958.
23. JAMES POLK8 BURSON (JAMES C.7, JOSEPH6, ISAAC5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1). He married HANNAH LOUISE (LULA )HUMPHRIES.
i. JAMES CODY9 BURSON, b. Coweta County, Georgia; d. Randolph county, Alabama.
ii. WILLIAM POLK BURSON, b. Coweta County, Georgia; d. Randolph County, Alabama.
iii. LILLER ANN BURSON, b. Randolphn county, Alabama; m. TOOMBS HESTER.
iv. WALTER CLEVELAND BURSON, b. Randolph county, Alabama; m. LUCILLE BENNETT.
55. v. LUCINDA PALESTINE (SIS) BURSON, b. July 23, 1881; d. 1973.
56. vi. HOMER JACKSON BURSON , WWI AND II, b. February 06, 1894.
24. NANCY8 YORK (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born December 25, 1827 in Darke Co., Ohio, USA. She married HENRY J. CRAIG January 13, 1853.
i. ALPHA9 CRAIG, b. 1855.
ii. JAMES CRAIG, b. 1863.
25. DELILA8 YORK (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born August 23, 1831 in Darke Co., Ohio, USA. She married (1) JOHN L. FERRELL March 17, 1856 in Mercer Co., Ohio, USA. She married (2) GEORGE PETRA 1867 in Mercer Co., Ohio, USA.
i. JOSEPH9 PETRA, b. 1855.
ii. THOMAS PETRA, b. 1860.
iii. BROADIS PETRA, b. 1867.
26. LOUISA8 YORK (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born October 19, 1834, and died April 23, 1878. She married JOHN HOLE June 21, 1855.
i. MARY9 HOLE, b. 1861.
ii. JOSEPH N. HOLE, b. 1863; d. 1892.
iii. ELIZABETH JANE HOLE, b. 1866; d. 1959.
iv. ADELIA HOLE, b. 1868.
v. JOHN C. HOLE, b. 1872; d. 1875.
vi. ELNORA J. HOLE, b. 1873; d. 1874.
vii. XURY HOLE, b. 1874; d. 1875.
27. ELMIRA8 YORK (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1840. She married JOSEPH COLLINS August 05, 1860.
i. LUCINA9 COLLINS, b. 1862.
ii. MARVIN GEORGE COLLINS, b. 1866.
28. REBECCA8 YORK (2) (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1842. She married ROBERT LARIMORE Bef. 1867.
i. ROBERT9 LARIMORE , JR., b. 1870.
29. EVAN8 YORK (JOSEPH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1848, and died 1937. He married MARIA HOUSEHOLDER August 17, 1871.
i. NEWTON NEWBERRY9 YORK, b. 1873.
30. MARY8 YORK (2) (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born December 22, 1830 in Darke Co., Ohio, USA, and died August 09, 1894 in Darke Co., Ohio, USA. She married STEPHEN G. HOLE December 23, 1849 in Darke Co., Ohio, USA.
i. PHOEBE ELLEN9 HOLE, b. 1850; d. 1873.
ii. LEWIS J. HOLE, b. 1853; d. 1931.
iii. MINERVA JANE HOLE, b. 1855; d. 1925.
iv. JEREMIAH HOLE (4), b. 1857; d. 1944.
v. HENRY HOLE (3), b. 1861; d. 1888.
vi. LEVI HOLE, b. 1862; d. 1943.
vii. MARY ELIZABETH HOLE, b. 1865; d. 1944.
viii. EMMA C. HOLE, b. 1867; d. 1924.
ix. LYDIA MARIAH HOLE, b. 1870; d. 1954.
31. LYDIA8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 07, 1832 in Darke Co., Ohio, USA. She married BENJAMIN C. MENDENHALL June 09, 1850 in Darke Co., Ohio, USA.
i. NANCY9 MENDENHALL, b. 1850.
ii. JOHN MENDENHALL, b. 1852.
iii. WILLIAM MENDENHALL, b. 1853.
iv. JOSEPH MENDENHALL, b. 1857.
v. BABE MENDENHALL, b. 1860.
32. CATHERINE ANN8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born March 31, 1835 in Darke Co., Ohio, USA, and died December 28, 1855 in Darke Co., Ohio, USA. She married SAMUEL EARHART June 06, 1854 in Darke Co., Ohio, USA.
i. FRANCES M.9 EARHART, b. 1855; d. 1866.
33. JOHN G.8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 03, 1836 in Darke Co., Ohio, USA, and died November 08, 1908 in Darke Co., Ohio, USA. He married HETTIE COZIER July 05, 1863 in Darke Co., Ohio, USA.
i. LUCINDA ANN9 YORK, b. 1864; d. 1926.
ii. IDA MAY YORK, b. 1867; d. 1941.
iii. HARRIET CATHERINE YORK, b. 1869; d. 1939.
iv. LAURABELL YORK, b. 1871; d. 1929.
v. HARVEY YORK, b. 1872; d. 1961.
vi. JOHN E. YORK, b. 1872.
vii. MINERVA JANE YORK, b. 1874.
viii. LEWIS YORK (3), b. 1876; d. 1953.
34. ELIZA JANE8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 08, 1837 in Darke Co., Ohio, USA, and died May 28, 1858 in Darke Co., Ohio, USA. She married JOSEPH HOGAN (2) July 29, 1855 in Darke Co., Ohio, USA.
i. FANANAUS9 HOGAN, b. 1858.
35. CAROLINE8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 11, 1838 in Darke Co., Ohio, USA, and died March 29, 1916 in Darke Co., Ohio, USA. She married (1) GEORGE C. COTTRELL REV. June 29, 1857 in Darke Co., Ohio, USA. She married (2) GEORGE W. CALDWELL June 17, 1895 in Darke Co., Ohio, USA.
i. JEREMIAH9 COTTRELL, b. 1860; d. 1932.
ii. REBECCA E. COTTRELL, b. 1861; d. 1934.
iii. DANIEL COTTRELL, b. 1862; d. 1948.
iv. SAMUEL H. COTTRELL, b. 1866; d. 1948.
v. STEPHEN H. COTTRELL, b. 1867.
vi. MARY JANE COTTRELL, b. 1868; d. 1953.
vii. GEORGE COTTRELL, b. 1870.
viii. LAURA BELLE COTTRELL, b. 1872.
ix. MINNIE PAULINE COTTRELL, b. 1873.
x. CLARA BELLE COTTRELL, b. 1876; d. 1953.
xi. LYDIA ELLEN COTTRELL, b. 1878; d. 1960.
xii. SARAH ELIZABETH COTTRELL, b. 1879; d. 1918.
36. WILLIAM8 YORK (4) (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 02, 1851 in Wayne Twp., Darke Co., Ohio, USA15,16, and died December 13, 1940 in Muncie, Delaware Co., Indiana, USA. He married AMANDA MARY PITTSENBARGER April 22, 1873 in Darke Co., Ohio, USA.
ii. CHARLES PERRY YORK, b. 1874, Patterson Twp., Darke Co., Ohio, USA17,18; d. June 30, 1953, Muncie, Delaware Co., Indiana, USA19,20.
iii. HARRISON FRANKLIN YORK, b. 1876, Patterson Twp., Darke Co., Ohio, USA21,22; d. December 20, 1940, Muncie, Delaware Co., Indiana, USA23,24.
iv. B. YORK, b. 1878, Darke Co., Ohio, USA25,26; d. October 20, 1878, Darke Co., Ohio, USA27,28.
v. CORA ETTA YORK, b. 1880, Darke Co., Ohio, USA29,30; d. September 04, 1881, Darke Co., Ohio, USA31,32.
vi. PEARL BELLE YORK, b. 1882, Darke Co., Ohio, USA33,34; d. December 26, 1948, Muncie, Delaware Co., Indiana, USA35,36.
vii. JOHN YORK (3), b. 1884, Darke Co., Ohio, USA37,38; d. October 10, 1885, Darke Co., Ohio, USA39,40.
viii. WILLIAM H. YORK, b. 1886, Darke Co., Ohio, USA41,42; d. January 16, 1888, Darke Co., Ohio, USA43,44.
ix. FRED YORK, b. March 1903, Muncie, Delaware Co., Indiana, USA45,46; d. 1915, Muncie, Delaware Co., Indiana, USA47,48.
37. RACHAEL8 YORK (JEREMIAH7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 27, 1856 in Wayne Twp., Darke Co., Ohio, USA49,50, and died March 02, 1932 in Versailles, Wayne Twp., Darke Co., Ohio, USA. She married (1) WILLIAM HENRY PITTSENBARGER October 03, 1872 in Darke Co., Ohio, USA. She married (2) WILLIAM GEORGE CYPHERS January 30, 1924 in Versailles, Wayne Twp., Darke Co., Ohio, USA51,52.
i. LULA VIOLA9 PITTSENBARGER, b. December 29, 1873, Willowdell, Patterson Twp., Darke Co., Ohio, USA53,54; d. June 09, 1953, Muncie, Delaware Co., Indiana, USA55,56.
ii. HENRY PITTSENBARGER, b. August 24, 1875, Willowdell, Patterson Twp., Darke Co., Ohio, USA57,58; d. August 25, 1875, Willowdell, Patterson Twp., Darke Co., Ohio, USA59,60.
iii. STEPHEN EDWARD PITTSENBARGER, b. October 28, 1876, Willowdell, Patterson Twp., Darke Co., Ohio, USA61,62; d. July 23, 1930, Versailles, Wayne Twp., Darke Co., Ohio, USA63,64.
iv. LOLA DELL PITTSENBARGER, b. September 27, 1878, Willowdell, Patterson Twp., Darke Co., Ohio, USA65,66; d. September 20, 1959, Muncie, Delaware Co., Indiana, USA67,68.
v. MARY JANE PITTSENBARGER, b. July 01, 1881, Versailles, Wayne Twp., Darke Co., Ohio, USA69,70; d. January 05, 1972, Sacramento, Sacramento Co., California, USA71,72.
vi. WILLIAM CAREY "BUD" PITTSENBARGER, b. April 18, 1883, Versailles, Wayne Twp., Darke Co., Ohio, USA73,74; d. June 09, 1963, Versailles, Wayne Twp., Darke Co., Ohio, USA75,76.
vii. AMANDA ELLEN PITTSENBARGER, b. August 30, 1887, Versailles, Wayne Twp., Darke Co., Ohio, USA77,78; d. September 19, 1965, Perrysburg, Wood Co., Ohio, USA79,80.
viii. CORA FRANCES PITTSENBARGER, b. January 26, 1890, Versailles, Wayne Twp., Darke Co., Ohio, USA81,82; d. July 31, 1963, Birmingham, Oakland Co., Michigan, USA83,84.
ix. RICHARD HARLAN PITTSENBARGER, b. July 24, 1893, Versailles, Wayne Twp., Darke Co., Ohio, USA85,86; d. January 03, 1952, Versailles, Wayne Twp., Darke Co., Ohio, USA87,88.
38. JOSEPH SLADE8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 10, 1835 in Darke Co., Ohio, USA, and died July 22, 1929 in Darke Co., Ohio, USA. He married (1) CATHERINE HULDA SHOOK July 05, 1855 in Darke Co., Ohio, USA. He married (2) RACHEL REAGON September 13, 1906 in Darke Co., Ohio, USA.
i. WILLIAM C.9 YORK, b. July 12, 1856, York Twp., Darke Co., Ohio, USA; d. February 25, 1857, York Twp., Darke Co., Ohio, USA.
ii. MARY JANE YORK, b. April 14, 1858, York Twp., Darke Co., Ohio, USA; d. March 1860, York Twp., Darke Co., Ohio, USA.
iii. ELIZABETH ALMEDA YORK, b. March 14, 1860, York Twp., Darke Co., Ohio, USA; d. December 08, 1886, Darke Co., Ohio, USA.
iv. NANCY ALDORA YORK, b. March 23, 1862, Darke Co., Ohio, USA; d. July 25, 1940, Darke Co., Ohio, USA.
v. NICHOLAS ALLEN YORK, b. March 03, 1864, Darke Co., Ohio, USA; d. April 09, 1933, Loramie Twp., Shelby Co., Ohio, USA.
vi. SARAH BELLE YORK, b. February 08, 1866, Darke Co., Ohio, USA; d. October 19, 1942, Darke Co., Ohio, USA.
vii. LAURA ELLEN YORK, b. May 04, 1868, Brock, York Twp., Darke Co., Ohio, USA; d. April 11, 1953, Greenville, Greenville Twp., Darke Co., Ohio, USA.
viii. PETER CALVIN YORK, b. December 03, 1870, Darke Co., Ohio, USA; d. December 13, 1870, Darke Co., Ohio, USA.
ix. JOSEPH FRANKLIN YORK, b. December 04, 1871, Darke Co., Ohio, USA; d. 1937, Darke Co., Ohio, USA.
x. MARTHA ANN YORK, b. February 23, 1874, Darke Co., Ohio, USA.
xi. MATTIE YORK, b. January 22, 1875, Darke Co., Ohio, USA; d. January 21, 1960.
xii. LULU ARZILPHA YORK, b. August 14, 1877, Darke Co., Ohio, USA; d. April 16, 1962.
xiii. HARVEY DELMON YORK, b. August 26, 1879, York Twp., Darke Co., Ohio, USA; d. April 15, 1952.
xiv. JOHN YORK, b. August 22, 1881, Darke Co., Ohio, USA; d. 1948.
39. SARAH JANE8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 06, 1837 in Darke Co., Ohio, USA, and died November 17, 1911 in Ohio, USA. She married WASHINGTON BAYMAN October 12, 1856 in Darke Co., Ohio, USA.
i. MARTHA ANN9 BAYMAN, b. December 12, 1857; d. July 05, 1888.
ii. IRVIN BAYMAN, b. December 08, 1859.
iii. NICHOLAS RILEY BAYMAN, b. April 08, 1862.
iv. UNKNOWN BAYMAN, b. 1864; d. 1864.
v. THOMAS E.BAYMAN, b. July 14, 1865.
vi. IDELIA BAYMAN, b. September 13, 1867; d. August 19, 1868.
vii. FRANKLIN BAYMAN, b. May 30, 1869.
viii. CHARLES BAYMAN, b. December 06, 1871.
ix. JOHN C. BAYMAN, b. March 16, 1874.
x. PEARL M. BAYMAN, b. September 26, 1877.
40. ELIJAH8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 26, 1840 in Darke Co., Ohio, USA, and died December 01, 1868 in Darke Co., Ohio, USA. He married NANCY ANN LEHMAN October 27, 1859 in Darke Co., Ohio, USA.
i. PETER FANKLIN9 YORK, b. May 16, 1860; d. 1934.
ii. JOHN CALVIN YORK, b. October 08, 1862; d. 1938.
iii. JEMIMA ELIZABETH "BETTY" YORK, b. July 18, 1864; d. 1951.
iv. SARAH JANE "JENNIE" YORK, b. September 04, 1866; d. 1936.
v. ELIJAH YORK (2), b. February 13, 1869; d. 1945.
41. IRVIN8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born September 26, 1842 in Darke Co., Ohio, USA, and died March 28, 1910 in Darke Co., Ohio, USA. He married SARAH ANN WINBIGLER April 21, 1862 in Darke Co., Ohio, USA.
i. CHARLES F.9 YORK, b. August 04, 1866; d. 1960.
ii. GEORGE S. YORK, b. April 08, 1868; d. 1959.
iii. DAISY LOUELLEN YORK, b. April 29, 1881; d. 1966.
42. DAVID8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born October 18, 1844 in Darke Co., Ohio, USA, and died October 05, 1925 in Darke Co., Ohio, USA. He married (1) CATHERINE HUGHES December 31, 1863 in Darke Co., Ohio, USA. He married (2) NELLIE MARTIN Aft. 1916 in Darke Co., Ohio, USA.
i. AWLIDA9 YORK, b. April 09, 1865; d. 1938.
ii. JEMINA ELZORA "BETTY" YORK, b. August 09, 1869; d. 1927.
iii. CHARLES S. YORK, b. May 16, 1880; d. 1881.
iv. ORPHA J. YORK, b. June 08, 1884; d. 1965.
v. DAVID LINDON YORK, b. March 15, 1886; d. 1886.
43. WILLIAM CALVIN8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born August 15, 1850 in Darke Co., Ohio, USA, and died January 20, 1935 in Shelbyville, Shelby Co., Indiana, USA. He married (1) NANCY ANN BOWMAN November 18, 1875. He married (2) CATHERINE MCDONALD March 04, 1886.
i. MINNIE9 YORK, b. 1886; d. 1948.
ii. JOSEPH FRANKLIN YORK (2), b. 1887.
iii. JAMES YORK, b. 1888; d. 1952.
iv. AGNES YORK, b. 1892; d. 1967.
v. THOMAS YORK, b. 1895.
vi. SUSANNA YORK, b. 1898; d. 1954.
44. NEWBERRY FRANKLIN8 YORK (NICHOLAS SLADE (ESQ.)7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born January 10, 1857 in Darke Co., Ohio, USA, and died October 21, 1921 in Ansonia, Brown Twp., Darke Co., Ohio, USA. He married MARY ANN LYONS September 21, 1875 in Darke Co., Ohio, USA.
i. JOHN LEONARD9 YORK, b. September 30, 1877; d. 1930.
ii. IRA YORK, b. June 22, 1879; d. Aft. June 22, 1879.
iii. SUSAN ANNA YORK, b. October 17, 1880, Darke Co., Ohio, USA; d. December 20, 1967, Darke Co., Ohio, USA.
iv. JEMIMA YORK, b. December 13, 1882; d. October 15, 1890.
v. BERTHA MAE YORK, b. June 18, 1885; d. 1984.
vi. GROVER CHARLES YORK, b. May 04, 1888; d. June 18, 1888.
vii. NORA J. YORK, b. July 25, 1889; d. 1967.
viii. ORANGE FREDRICK YORK, b. February 22, 1892, Ohio, USA; d. 1979.
ix. FLORENCE LUVERNA YORK, b. August 13, 1895; d. 1988.
x. RAYMOND LEE YORK, b. September 08, 1898; d. 1923.
xi. THOMAS N. YORK, b. May 08, 1903.
45. NEWBERRY F.8 MILLER (DIADAMA7 YORK, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 20, 1840, and died September 21, 1921. He married HETTIE ANN MARKER October 17, 1869 in Cedar Co., Iowa, USA.
46. ROSETTA8 YORK (WILLIAM7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 12, 1830, and died October 11, 1877. She married JACOB B. BRUNS Abt. 1853.
i. NICHOLAS C.9 BRUNS, b. 1854; d. 1936.
ii. DANIEL L. BRUNS, b. 1857.
iii. NANCY A. BRUNS, b. 1862.
iv. JACOB CALVIN BRUNS, b. 1865; d. 1945.
47. NANCY8 YORK (2) (WILLIAM7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born September 21, 1833, and died February 07, 1865. She married SAMUEL DAY November 21, 1861.
i. WILLIAM F.9 DAY, b. 1863; d. 1939.
ii. MARIA DAY, b. 1865.
48. NICHOLAS J.8 YORK (WILLIAM7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born Abt. 1838, and died 1908. He married CATHERINE A. DUNCAN November 09, 1865.
i. JAMES D.9 YORK, b. 1874; d. 1897.
ii. MINNIE E. YORK, b. 1878.
iii. WILBUR YORK, b. 1880.
49. EFFA8 YORK (WILLIAM7, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1844. She married ANDREW WATTERLY December 31, 1868.
i. JESSE9 WATTERLY, b. 1870.
ii. LAURA WATTERLY, b. 1872.
iii. CHARLES WATTERLY, b. 1875; d. 1966.
50. JAMES D.8 ARMSTRONG (REBECCA7 YORK, NEWBERRY6, LYDIA5 BURSON, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born 1848 in Ohio, USA, and died 1927. He married MARGARET NEARGARDNER January 20, 1869 in Darke Co., Ohio, USA.
i. SAMUEL9 ARMSTRONG (2), b. 1871, Ohio, USA.
ii. BARBARA ARMSTRONG, b. April 30, 1873, Darke Co., Ohio, USA; d. April 20, 1875, Darke Co., Ohio, USA.
iii. MARIA ELIZABETH ARMSTRONG, b. April 15, 1875, Darke Co., Ohio, USA; d. 1880.
iv. DAVID DELROY ARMSTRONG, b. January 02, 1879, Frenchtown, Wayne Twp., Darke Co., Ohio, USA; d. 1960.
v. JOHN ARMSTRONG, b. June 22, 1881, Darke Co., Ohio, USA; d. November 12, 1965, Darke Co., Ohio, USA.
vi. CHARLES ARMSTRONG, b. June 01, 1891, Darke Co., Ohio, USA; d. June 30, 1964, Dayton, Harrison Twp., Montgomery Co., Ohio, USA.
vii. LINUS ARMSTRONG, b. September 19, 1894; d. May 23, 1951, Darke Co., Ohio, USA.
viii. UNKNOWN ARMSTRONG, b. June 22, 1896, Darke Co., Ohio, USA; d. June 22, 1896, Darke Co., Ohio, USA.
ix. NORA ARMSTRONG, b. 1898; d. 1898.
x. ALPHA L. ARMSTRONG, b. 1901.
xi. EMMA FLORENCE ARMSTRONG, b. 1904.
51. LULA BELL8 HURST (MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born September 27, 1861, and died May 31, 1929. She married KYSER COX December 21, 1861.
57. i. ANNIE BLANCH9 COX, b. September 16, 1880; d. February 11, 1914.
58. ii. WILLIE MAUDE COX, b. February 08, 1883.
59. iii. CLARENCE KYSER COX, b. May 24, 1887.
52. WILLIAM JAMES JACKSON8 HURST (MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born December 22, 1863. He married IDA MABEL SEALS.
60. i. MITTIE BELL9 HURST, b. April 04, 1885.
ii. WILLIE SEALE HURST, b. May 10, 1887; m. LULA ROBSON, December 31, 1910.
iii. ELLIE BERTHA HURST, b. July 07, 1889; m. CHARLES SMITH, November 29, 1924.
iv. MARY LOLA HURST, b. July 26, 1891; m. WALTER DENTON, October 18, 1823.
v. JOHN STANFORD HURST, b. February 18, 1893; d. October 18, 1938; m. EVA DOUGLAS.
61. vi. MAXINE LULA HURST, b. February 04, 1895.
vii. MATTIE MAE HURST, b. January 17, 1897; m. CHARLES STOCKDALE.
62. viii. MATTIE GUTHRIE HURST, b. March 24, 1899.
63. ix. MIRIAM HURST HURST, b. April 05, 1903.
53. ELKANAH LAWRENCE EARLY8 HURST (MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 09, 1867. He married TOMMIE DE ARMAN.
64. i. DICK9 HURST, b. November 02, 1891.
ii. RAYMOND HURST, b. September 21, 1894, Baggett, Texas; d. February 05, 1918.
54. ELKANAH GEORGE8 BURSON DR. (ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 07, 1882 in Furman, Wilcox County, Alabama89, and died April 1970 in Furman, Wilcox County, Alabama buried in Palmer Cemetery90. He married ELIZABETH JANE KNIGHT February 20, 1914 in Selma, Alabama, daughter of NAPOLEON KNIGHT and MARGARET WATSON.
65. i. ELLIECE9 BURSON, b. April 07, 1915, Furman, Alabama, Wilcox County; d. 1994, Selma, Alabama, buried Palmer cemetery.
66. ii. ELKANAH GEORGE BURSON , DR, b. July 04, 1918, Furman, Wilcox County, Alabama.
iii. ELIZABETH BURSON, b. March 11, 1920; m. JAMES DUDLEY HART, April 23, 1949.
55. LUCINDA PALESTINE (SIS)9 BURSON (JAMES POLK8, JAMES C.7, JOSEPH6, ISAAC5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 23, 1881, and died 1973. She married WILLIAM ASBERY MOSES.
56. HOMER JACKSON9 BURSON , WWI AND II (JAMES POLK8, JAMES C.7, JOSEPH6, ISAAC5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 06, 1894. He married (1) ERA PRESCOTT. He married (2) FLOSSIE CRAWFORD.
x. LENA BURSON, m. GREEN.
xiii. TOOMBS B. HESTER BURSON , WWI AND II.
57. ANNIE BLANCH9 COX (LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born September 16, 1880, and died February 11, 1914. She married DEKLE March 28, 1931.
67. i. LULA BLANCH10 DEKLE, b. October 18, 1913.
58. WILLIE MAUDE9 COX (LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 08, 1883. She married COURTLAND PRESTON ALYEA September 07, 1903.
68. i. COURTLAND PRENTISS10 ALYEA, b. June 03, 1904.
69. ii. CLIFFORD PRESTON ALYEA, b. November 14, 1906.
59. CLARENCE KYSER9 COX (LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 24, 1887. He married MARGARET ADELLE MAY December 31, 1907.
70. i. VIVIAN KYSER10 COX, b. October 31, 1908.
ii. MARGARET BLANCH COX, b. September 01, 1914.
60. MITTIE BELL9 HURST (WILLIAM JAMES JACKSON8, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 04, 1885. She married WILIE GRIMES May 24, 1906.
i. WYOLA10 GRIMES, b. March 24, 1907; m. EDGAR PIERCE, March 24, 1928.
ii. WILEY HURST GRIMES, b. September 17, 1908; m. BONNE MAE GOODSON.
iii. WILLIAM MILFORD GRIMES, b. May 06, 1911.
iv. MILDRED GRIMES, b. April 18, 1913; m. JAMES CLYDE TIREY.
61. MAXINE LULA9 HURST (WILLIAM JAMES JACKSON8, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born February 04, 1895. She married EDGAR MILTON June 26, 1916.
i. JOHN EDGAR10 MILTON, b. October 21, 1924.
62. MATTIE GUTHRIE9 HURST (WILLIAM JAMES JACKSON8, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born March 24, 1899. She married CHARLES STOCKDALE.
i. WALTER10 STOCKDALE, b. June 13, 1916.
63. MIRIAM HURST9 HURST (WILLIAM JAMES JACKSON8, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 05, 1903. She married HENRY JAMES STABLER.
i. MIRIAM10 STABLER, b. August 06, 1922; m. JAMES IRA TILL, June 17, 1940.
ii. WILLIAM LAMAR STABLER, b. August 29, 1924.
iii. IDA MARY STABLER, b. February 17, 1928.
iv. HENRY FRANCIS STABLER, b. June 15, 1936.
64. DICK9 HURST (ELKANAH LAWRENCE EARLY8, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 02, 1891. He married CLARA BELL CHILDS August 25, 1921.
i. DOROTHY ESTHER10 HURST, b. April 26, 1931.
65. ELLIECE9 BURSON (ELKANAH GEORGE8, ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 07, 1915 in Furman, Alabama, Wilcox County, and died 1994 in Selma, Alabama, buried Palmer cemetery. She married (1) EXTON TUCKER. She married (2) CLAUDE WILLIAMS September 18, 1933.
71. i. CLAUDE BURSON10 WILLIAMS, b. January 04, 1935, Greenville, Alabama, Wilcox County; d. Madisonville, Kentucky.
66. ELKANAH GEORGE9 BURSON , DR (ELKANAH GEORGE8, ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born July 04, 1918 in Furman, Wilcox County, Alabama. He married JEAN BRONSON GILLIS August 03, 1947 in Brewton, Alabama Escambia County, daughter of JOHN GILLIS and CLARE JERNIGAN.
72. i. SHARMAN JEAN10 BURSON, b. May 15, 1950, Dothan, Houston County, Alabama.
ii. SYLVIA JANE BURSON , DR., b. July 31, 1956, Dothan, Houston County, Alabama; m. (1) RAMY TAWFIK; m. (2) FRANK BAROVECCHIO, 1976, Dothan, Al.; m. (3) THOMAS JEROME RUSHING, December 15, 1989, New Orleans, Louisiana.
73. iii. ELKANAH GEORGE BURSON III, b. April 12, 1960, Dothan, Al..
67. LULA BLANCH10 DEKLE (ANNIE BLANCH9 COX, LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born October 18, 1913. She married CHARLES WESLEY MILLER , JR..
i. CHARLENE K.11 MILLER, b. November 16, 1941.
68. COURTLAND PRENTISS10 ALYEA (WILLIE MAUDE9 COX, LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born June 03, 1904. He married ANNA EICKLES August 1925.
i. ANNA BELL11 ALYEA, b. April 15, 1928.
69. CLIFFORD PRESTON10 ALYEA (WILLIE MAUDE9 COX, LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born November 14, 1906. He married RUTH PEARL HARDESTY June 23, 1928.
i. CLIFFORD PRESTON11 ALYEA, b. February 05, 1929.
ii. MARLENE RUTH ALYEA, b. October 18, 1932.
iii. JAMES HARVEY ALYEA, b. September 05, 1939.
70. VIVIAN KYSER10 COX (CLARENCE KYSER9, LULA BELL8 HURST, MARTHA ANN7 BURSON, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born October 31, 1908. He married DANIEL EVELYN ALLISON October 16, 1937.
i. ANNYAN ALLISON11 COX, b. 1941.
71. CLAUDE BURSON10 WILLIAMS (ELLIECE9 BURSON, ELKANAH GEORGE8, ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born January 04, 1935 in Greenville, Alabama, Wilcox County, and died in Madisonville, Kentucky. He married (1) DORCAS ANNETTE (TOBY) BOYD August 17, 1958 in Chattanooga, Tennessee. He married (2) DEBRA DILLEY 1977 in Henderson, Kentucky.
i. HAYDEN EUGENE11 WILLIAMS, b. March 09, 1961, Madisonville, Kentucky; m. KERRIE.
ii. CLAUDE BURSON WILLIAMS, b. March 09, 1961, Madisonville, Kentucky; d. November 19, 1996, Kentucky; m. (1) LISA; m. (2) PATTY.
iii. BLAKE WILLIAMS, b. June 20, 1964.
iv. CHRISTOPHER RYAN11 WILLIAMS, b. May 22, 1978.
72. SHARMAN JEAN10 BURSON (ELKANAH GEORGE9, ELKANAH GEORGE8, ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born May 15, 1950 in Dothan, Houston County, Alabama. She married JOEL WARDLAW RAMSEY, son of JOSEPH RAMSEY and HILDA HAWKINS.
i. CECILY CATHRYN11 RAMSEY, b. May 02, 1974, Dothan, Alabama, Houston County; m. STEPHAN ALLAN BUTTERWORTH, May 05, 2001, 17807 Hwy. 98, Panama City Beach, Florida.
ii. ANDREW ALLEN RAMSEY, b. March 13, 1976, Dothan, Alabama Houston County.
iii. BETHANY BROOKE RAMSEY, b. May 20, 1980, Dothan, Alabama Houston County.
73. ELKANAH GEORGE10 BURSON III (ELKANAH GEORGE9, ELKANAH GEORGE8, ELKANAH7, JOSEPH (JACKSON)6, JOSEPH JEFFERSON5, JOSEPH4, JOSEPH3, GEORGE2, GEORGE1) was born April 12, 1960 in Dothan, Al.. He married DEBRAH DEAN BUFORD February 08, 1987 in Tuscaloosa, Alabama.
i. HANNAH ELIZABETH JEAN11 BURSON, b. March 07, 1991.
ii. REBECCA JEWEL BURSON, b. November 01, 1994, Mobile, Alabama.
iii. MALLORY MARTHA JANE BURSON, b. September 04, 1996.
iv. MARY ELLE BURSON, b. February 1999.
| 2019-04-20T22:14:17 |
http://southern-style.com/Burson%20Genealogy.htm
|
0.999973 |
Why are there so many Islands?
I am currently studying geology on the island of Santorini. One of the concepts that is really important in doing that is how the islands in the Cyclades form. For most of the islands, plate tectonics hold the answer.
The idea of plate tectonics is that the Earth’s crust is made up of plates. These plates move around and crash into each other which is what causes most of Earth’s volcanism and earthquakes.
At the edges of each of these plates are smaller plates that are called microplates. These are usually grouped with the larger plates because they behave in a similar fashion. However, in the Aegean, it is important to understand that these plates are separated from the main plate by fault lines and sometimes act separate from the main plate.
In the Mediterranean, the African plate in moving under the Eurasian plate and the Aegean microplate. When one plate moves under another plate. It is called subduction. This creates a trench in the sea floor at the plate boundary which is marked in red in the above map. Parallel to subduction zones are volcanic arcs, including one within the Aegean Sea.
As the subducting plate sinks into the Earth, it pulls the overriding plate with it. This is known as subduction rollback and it makes the overriding plate move faster near the plate boundary.
The pulling from the subduction rollback creates uneven movement within the overriding plate. This leads to extension and thinning of the crust within the plate or microplate.
In order to thin the crust, normal faults form. Faults are breaks in the earth. Normal faults have a hanging wall and a footwall. The hanging wall is on top of the footwall. In a normal fault the hanging wall moves down relative to the footwall. The faults in the Aegean tend to point in a line from northeast to southwest.
As the hanging wall moves down relative to the footwall in a normal fault, the footwall also moves up relative to the hanging wall. After enough time has passed and the footwall has moved enough, the footwall will emerge above the water as an island. This is how most of the islands in the Cyclades are formed.
However, a few of the islands were constructed by volcanism and are either active or dormant volcanoes. These volcanoes are located on a different type of fault called a strike-slip fault. Strike-slip faults do not have a hanging wall and a footwall, but move horizontally whereas normal faults move vertically.
The strike-slip faults that are located with the volcanoes are special in that they are oblique strike-slip faults, so they move predominately horizontally but also move vertically. This type of motion creates a gap where the magma can move to the surface. Which is why there are volcanoes on these strike-slip faults, but no volcanoes along a strike-slip fault such as the San Andreas.
There are only four volcanic centers within the Aegean Sea. The rest of the islands all formed from normal faults. The four centers in order from West to East are Methana, Milos, Santorini, and Kos/Nysiros. These are circled in pink on the map of the Aegean above.
Plate tectonics created faulting, which then created islands and allowed magma to flow through the crust and make volcanic islands. I think this is a very important concept to understand in order to truly comprehend why there are so many islands within the Aegean Sea.
You have a good sense of organization, and you have a good impulse for which kinds of visuals to include. I hadn’t learned about strike-slip faults before, so I find this post very interesting. In fact, this text is intriguing and caught my attention.
I read it over 3x and wished you had included an image or two. In fact, when you’re teaching, as you are in this post, include visuals that keep us moving along with your discussion. You could, for instance draw the differences between the strike-slip faults. Take advantage of your field notebooks and use columns, cross-sections, overheads, and close-ups.
It’s clear that you understand quite a bit and perhaps, your daily experience among other Geologically minded people makes it difficult to write for the rest of us. It’s a good idea to find a blog-buddy or three, and trade drafts before posting. Write for someone you know who is like me—very interested in your course, in Santorini, and I look forward to learning from your experiences and new knowledge.
A place called the Hungry Donkey, well, take pictures.
I look forward to your next post.
You have a detailed post with lots of information, but for someone who has little to no background in Geology, they may have a difficult time picturing and understanding what you are describing. This would be a great opportunity to let your voice come through, use your experiences, diagrams, and even some everyday comparisons that may help breakdown the more complicated Geology lingo.
Agree with both Sheridan and Professor Barron that you have a good understanding of the material, but this post is a little lingo heavy. It would be great, for example, to include a diagram of normal faulting explaining the hanging wall and footwall. Also, adding the stresses to your cross section (where extension is happening). The maps are well chosen, but they would benefit from some annotation. Where all all the islands you are talking about, etc.
| 2019-04-21T02:43:32 |
https://nauingreece.blog/2018/06/08/why-are-there-so-many-islands/
|
0.999563 |
The Wave-Disk engine. What? I don't know, but if they can have a prototype by the end of the year as promised, this is a whole new direction for cars.
The Wave Disk Generator uses 60 percent of its fuel for propulsion; standard car engines use just 15 percent. As a result, the generator is 3.5 times more fuel efficient than typical combustion engines.
Researchers estimate the new model could shave almost 1,000 pounds off a car's weight currently taken up by conventional engine systems.
I would never buy a car with out it having a Marshall Engine.
| 2019-04-18T10:26:57 |
http://hectorowen.blogspot.com/2011/04/maybe-that-new-car-should-have-one-of.html
|
0.998292 |
The good thing about internet marketing is that is helps align on the way that the consumers will make their decisions in terms of buying. Increasing number of consumers make use of the social medial and research on the mobile internet in order for them to carry out their initial products and price research prior to making final decision. The good thing about internet marketing is that it helps in building relations with the customers through regular personalized communication, in reflection to the move away to that of the marketing.
By the help of the internet marketing, you can overcome the barriers of distance. You can sell different goods in any part of the country without having your local outlets thus widening your market target. You can also have an export business even without you opening a network for the distributors in various parts of countries. But if ever you want to sell internationally, you must use the localization services in order to ensure that the products are best suited for the local markets and comply to that of the local business regulations. The localization services will include those product modification and the translation in order to reflect the differences in the local market.
Moreover, if you market the products into the internet it will cost less than marketing it to the physical retail outlets stores. Obviously, you will not have the recurring costs of rental property and for its maintenance. The good thing is that you do not have to purchase for stocks in your store so you do not have to display something. It will keep the inventory costs low by ordering stock in line with that of the demand.
Social media can surely benefit from the internet marketing. There is an increased in sales among group of consumers that had responded to the influence of the social networks. You can also benefit in this type of influence by simply incorporating it to the social networking tools towards the campaign of internet marketing. Lastly, you can surely benefit from the internet marketing since it will lessen the worry about store opening and paying the staff for their overtime. It can also convenient for the customers especially in offering the products. Clients now can simply browse in the internet if they want to buy something in the online store at any of the given time and place that will be very convenient for them when they are just in their home and do not want to go shopping outside. Internet marketing will be a great avenue for many to sell their products.
| 2019-04-19T00:44:38 |
https://articledirectoryhere.com/figuring-out-marketing/
|
0.999971 |
Let's picture the following situation - let's say, you're on your way to an important business meeting, you stop by for some coffee, go back to your Kia and realize - "I locked my keys in my car!" Or, coming back to a parking lot with shopping bags and realizing you dropped your car keys somewhere. These situations usually happen when people are in a hurry. Stress just adds up, and you need a quick solution to the problem. But, who can solve your car lock-related issue and how?
To solve this type of situation, your best decision would be to contact us - the best auto unlock service! It is always better to hire a professional to take care of your car-lock related snag quickly and professionally than to do something irrational and damage your Kia. Turning to certified experts, like ourselves, means paying a little money, but getting an outstanding service.
Professional training and equipment - when you contact our locksmith service, you can be sure to be sent an employee with a proper unlock car door kit for your Kia. Our technicians have also gone through a necessary training on how to do this type of job to get it done fast, and never do any harm to your vehicle. Also, if you tried to open your car door by yourself and damaged it, they can do a car door lock repair. They are always fully equipped with the necessary tools.
It's normal that a cost to unlock Kia car door can vary depending on the locksmith businesses. Not every company values their work the same, but one of the biggest problems nowadays is that more and more people have complained about being tricked into paying a lot of money for an auto unlock service.
"I left my keys in my car and its locked. What should I do now, for God's sake? It's 3 a.m.!" This is real life, and we know it. That is why you can count on us 24 hours a day, seven days a week, year round. We know how important it is to have someone to help you when you get stranded somewhere, alone, in the middle of the night. Auto locksmith service is doing a good job only when it's available 24/7. That is why we are here to perform an emergency lockout on your Kia and to bring back the safety whenever you need us. After we receive your call, we immediately send out a skilled technician your way.
Having keys locked in car can be a nightmare, but if you find professionals like us to solve it for you, it can be something you resolved quickly, and even turn into a pleasant experience (since our staff is very polite). Let us help you and restore the security to your Kia in the shortest possible time frame.
If you've got locked keys in your Kia, call us right away at 888-870-3612, and we will help you! Available 24/7, licensed, very affordable, No.1 in town!
| 2019-04-24T12:16:05 |
http://www.auto-locksmith.com/keys-locked-in-car.asp?Kia
|
0.99997 |
The World Wide Web came about in the 1990's. In the 2000's, the preponderance of "blogs" (short for "weblogs"), as opposed to traditional websites, changed the feel of entire web.
To answer that question, one must first understand what a "blog" is. What is it that differentiates a blog from a traditional website? What is it that has made the blog so successful?
The reason people made so much fuss about the world wide web when it first appeared is that websites allow absolutely anyone to broadcast information easily to hundreds of millions of people all over the world, in an ever-growing variety of media. Websites blurred the boundary between the active broadcaster of information (i.e. a person or organization with money) and the passive consumer of information. The level of change is comparable to that experienced in Europe when the printing press was invented and manuscripts no longer had to be copied by hand by people with infinite time (i.e. employed by the powerful Catholic church).
In the early days of the web, people's websites sometimes had "news" or "recent changes" section that the sites' maintainers updated on a semi-regular basis. Soon, a few people started to use these "recent changes" sections to start writing about their lives. Online diaries appeared. Programs were written to make it easier for people to write periodic site updates without having to manually edit their HTML files. Eventually sites like Livejournal and Blogspot allowed people to broadcast regular updates on their lives without even owning a website.
Essentially, blog entries are like speeches you make to friends: they are meant to be posted, discussed, and eventually replaced, in an ever-evolving discussion. Starting a blog is not much more difficult than calling up a group of your friends and ranting at them. You don't even have to bother with setting up a website to own a popular, well-trafficked blog. All you need is a computer with an Internet connection.
The first step to starting your own blog is to find a blogging platform that you want to use. Sites like blogger.com and wordpress.com are among the most popular of the blogging platforms used by new bloggers because they are free, easy to use, and have lots of resources. If you have an e-mail address (which is also easy to obtain for free), you can set up an account with one of these sites and start writing entries within minutes. A slightly more complicated way to start a blog, if you already have a traditional web site and are already hosting your own content, is to download useful software from wordpress.com or movabletype.org. That way, you can put a Wordpress or Movabletype-powered blog on a site that already has your own domain name (otherwise, your blog's URL would have to be something like Bobsblog.blogspot.com).
Of course, the trickiest part of starting a blog isn't the setting up. The trickiest parts of starting a blog are the twin complexities of writing for the blog, and promoting it. Presumably, you're starting a blog because you want your thoughts to be heard by millions. Should you just start posting blog entries about whatever comes into your head, and hope that someone will stumble upon them, perhaps from a random Google search? That's been done, especially in the early days of blogging.
Today, however, the Internet is very large. You'll be competing with thousands, if not millions, of other blogs for readers who have never heard of yours. You run the risk of no one, or close to no one reading or responding to your writing. This can be disappointing to the aspiring blogger.
As such, the next step after starting a blog is finding a theme for your blog. People no longer read blogs just to read blogs. They read them for specific reasons: for example, to keep informed about politics; or, to get arts and crafts ideas; or, to read someone's opinion about their local sports team; or, to get their "regular fix" of good short stories. You could have a blog that's just about your life, but, here, especially, you need to take particular care that it's interesting to other people to read. People don't want to read a random blog of what some random person does every day. They want to read a witty, thoughtful, idiosyncratic account of someone's particular, idiosyncratic life. Keeping your blog focused on a single theme or subject will make it easier for you to write--and will make it likelier to attract readers.
After you've found a theme you're interested in and have a lot to say about, start writing entries. At the same time, begin the first steps of blog promotion: talk about it to your friends. Find others who blog about the same subjects (sites like wordpress.com and blogger.com let you do this easily), and befriend them by commenting on their blogs. Write on message boards and forums that relate to what your blog is about (you can find these easily via Google search and clicking on links), and put a link to your blog at the end of each of your comments as a kind of signature. Create an identity for yourself on important forums and message boards, and in the minds of other important bloggers.
Advertise your website, business or product page on the "How to Blog" page and help support the free tips website!
| 2019-04-24T23:52:47 |
http://www.toptipspot.com/tips/computers/howto/blog.php
|
0.998745 |
35 + 7 – 5 ¸5 × 6 = ?
2 . One morning at 7 o’clock, Naresh started walking with his back towards the Sun. Then he turned towards left, walked straight and then turned towards right and walked straight. Then he again turned towards left. Now in which direction is he facing ?
3 . From the given alternatives select the word which can be formed using the letters of the given word.
4 . 3 daily wage workers A, B and C are distributed Rs.178 in such a way that A gets Rs.4 less than C, B gets Rs.15 more than A and C gets Rs.11 less than B. What is the ratio of their shares ?
5 . If the 5th date of a month is Tuesday, what date will be 3 days after the 3rd Friday in the month?
6 . Seema’s younger brother Sohan is older than Seeta. Sweta is younger than Deepti but elder than Seema. Who is the eldest ?
7 . Rahim and his uncle differ in their ages by 30 years. After 7 years, if the sum of their ages is 66, what will be the age of the uncle ?
8 . Sohan ranks seventh from the top and twenty-sixth from the bottom in a class. How many students are there in the class ?
| 2019-04-22T19:11:27 |
https://ssconlineexam.com/onlinetest/SSC-CGL-Tier-1/General-Intelligence-and-Reasoning/GIR-Test-10
|
0.999829 |
Is The Fur-Bearers a non-profit or a charity?
The Association for the Protection of Fur-Bearing Animals (The Fur-Bearers) is a registered non-profit organization (S-4222). Due to our committed and unapologetic advocacy for fur-bearing animals and wildlife, under Canada Revenue Agency regulations, we are not eligible for charitable status.
Though the terms "non-profit" and "charity" are often used interchangeably, a non-profit differs from a charity in a number of ways, most notably: a non-profit organization is not able to issue a tax receipt while a registered charity can. Registered charities are also eligible for considerably more grants and government funding than a non-profit. Many people ask why The Fur-Bearers does not have charitable status, and in order to answer this, a quick review of The Fur-Bearers history is needed.
Why can’t The Fur-Bearers have charitable status?
The Fur-Bearers was started in 1953 and actually began as a registered charity. The mandate was to expose the inherent cruelty associated with killing animals for the commercial fur trade.
In 1998, despite having had charitable status for decades, Revenue Canada (now the Canada Revenue Agency) contacted The Fur-Bearers stating that they “made a mistake” and that The Fur-Bearers should never have received charitable status in the first place. The Fur-Bearers was committed to showing the public the truth about the fur trade, however since charities in Canada cannot criticize a legal industry, our work was not considered a charitable endeavor. They also argued that we were spending more than the allowed 10% of our revenue on advocacy campaigns. Revenue Canada made it clear that unless The Fur-Bearers was willing to change its mandate, our charitable status would be “annulled”.
The Fur-Bearer's volunteer Board of Directors refused to be silenced, and believed that the organization’s work to expose the horrors of the fur trade was more important than being a charity. The Fur-Bearers carried on as usual, lost charitable status, and has continued ever since as a non-profit.
There were consequences to losing charitable status, primarily with funding, which was initially reduced by 50%. However, many supporters continued to support The Fur-Bearers, understanding that The Fur-Bearer's work was so critical they were willing to forgo a tax receipt. There were (and still are) benefits to being without charitable status in Canada.
Despite a loss in funding, without charitable status, we are now able to work harder than ever to push for an end to the commercial fur trade and meaningful legislation that protects fur-bearing animals and other wildlife. Every day we take full advantage of the opportunity afforded to us by being a non-profit (and not a charity).
For example, we work in municipalities across the country to end the use of bodygripping traps (leg-hold, Conibear, and snare) and have had great successes. We are fighting to ban the import and sale of dog and cat fur in Canada. We lobby municipal and provincial governments to end the trapping and culling of wildlife, and do so with great success. We encourage the development of laws at various levels of government that offer animals more protection. Much of this lifesaving work would not be possible if The Fur-Bearers had charitable status.
We are able to do all these things and more, thanks to the humbling generosity of individuals who care more about animals than they do tax receipts.
We thank each and every one of them for their continued support.
| 2019-04-22T02:51:24 |
https://thefurbearers.com/about-us/loss-of-charitable-status
|
0.998825 |
The government has reprieved five courts identified for closure as part of its review of the estate.
The Ministry of Justice said it intends to close 64 sites as proposed in its consultation last year.
A further 22 closures will take place but with changes to the original proposals. The full list of closures, which are intended to take place within the next two years, can be found here, with the timescale for each closure stated here.
The five courts to stay open will be St Helens County Court (although the magistrates’ court will close), Stockport, West Cumbria, Bath and Carmarthen – civil and family hearing centre, the law courts at the Guildhall (pictured) will close.
The government said each closure - affecting one-fifth of the total court estate in England and Wales - will still allow for 'effective access to justice' and the maintenance of 'high-quality service provision'.
Justice minister Shailesh Vara said: ‘Court closures are difficult decisions; local communities have strong allegiances to their local courts and I understand their concerns.
Vara said 97% of citizens will be able to reach their required court within an hour by car, even after the closures. The proportion able to reach a tribunal within an hour will be 83%.
While some campaigners fighting to keep their local court open will be relieved, many court users will be disappointed that so few sites were identified for reprieve. The Law Society had suggested there were good reasons to save 59 courts across England and Wales.
The consultation, which opened last July, prompted more than 2,100 separate responses, along with 13 petitions containing more than 10,000 signatures.
The government's response stressed that access to justice is 'not just about proximity to a court' and that changes to estimated travel times are 'small'.
The MoJ said in 'exceptional cases' start times of hearings may be delayed to allow later attendance due to travel difficulties.
However he welcomed that five of the proposed closures would not be going ahead and acknowledged that the government intends to make changes to its proposals for 22 courts following evidence submitted by our members. ‘We look forward to continuing discussions about the revised proposals in these areas to ensure that access to justice problems are mitigated.
The Public and Commercial Services Union, which represents court staff, said the government’s rationale for closures ‘misrepresents courts usage and journey times’ and the decision will make it harder for witnesses and victims to attend court.
General secretary Mark Serwotka said: ‘The closures would seriously undermine access to local courts that is crucial not only for the administration of justice but also public confidence in the justice system.
Jo Edwards, national chair of dispute resolution service Resolution, said the government has yet to present any analysis on the impact on courts that will remain open and absorb work from closed sites.
This is a disgrace. It has nothing to do with justice and is just about saving money.
Interestingly the print edition of the LSG has the headline "Five courts saved from closure" as if that was something to celebrate, which of course for those working in those 5 courts it is, but the wider picture is totally appalling when the vast majority of those on the list are closing.
Thank you Anon at 10.07 pm - there aren't many laugh-out-loud contributions to Readers' Comments, probably because the Gazette articles are so thoroughly depressing.
Note for editor: after right-clicking the picture and selecting 'inspect' I was able to establish that the picture is of Carmarthen Courts. Very attractive. Probably one of the nicest photos you've put up there in a while. Thinking of attractive photos - do you have another one of that lovely Elisabeth Davies?
"Or the time that the court bounced a charging order application because I'd tried to pay two application fees (there were two defendants as tenants in common). After arguing with them for ages I just sent them one fee, whereupon it was returned to me as apparently I needed two."
"Vara said 97% will be able to reach their required court by car within an hour".
However, if the government pushes ahead and bans whiplash claims it is not going to be safe for any of us to travel by car any more. And if you live in a rural area public transport probably isn't a viable alternative (even if you'll automatically get compo if the train is 30mins late).
I am just listening to the Today programme. There is a report from Hamburg. There is a protest against Polish immigrants in particular, and immigration in general. The extreme right Aktion fur Deutschland is making progress in the polls and elections as a result.
Not only will all this immigration test our infrastructure to breaking point, it will have political consequences too.
35.1 million cars on the road in the UK in 2015. Population of UK 64.1 million as at 2013. Even on the most basic "fag packet" calculations, the numbers don't add up. Or divide. Or something.
I'd be interested to see the one third stats. Some of the courts (for example Kettering) have never sat 5 days or even every week, but have operated as branch offices to process claims and as hearing centres and, more recently to take overspill cases from their centres. So they may sit 100% of their allotted time.
Once these courts are closed, I presume that listing rooms will be provided for the cases migrating to the remaining courts. Since, even if there are enough judges (debateable) they will be no use unless they have somewhere to do the judging.
I particularly like having an interim charging order application returned from X County Court with a covering letter from X County Court telling me that the case has been transferred to X County Court. That particular gaffe let the defendant sell their property before the interim order was made.
I also like being able to get all the way to trial without the defendant ever having filed an actual defence (by which I mean, the 'I want to defend the claim' box was ticked but no reasons filled in) because the court never got around to considering any strike-out or summary judgment requests.
Or that time it took me 2 years (no exaggeration) to get judgment in default against a defendant, because the County Court kept losing the paperwork. In the end I had to apply for summary judgment (which, interestingly, shouldn't have been possible).
Or the time that the court bounced a charging order application because I'd tried to pay two application fees (there were two defendants as tenants in common). After arguing with them for ages I just sent them one fee, whereupon it was returned to me as apparently I needed two.
Or about 50% of the time that a request for default judgment with contractual interest is bounced because the interest claimed is not the court rate of 8%.
Or at the court counter where I was told I needed 5 copies of a consent order to be stamped. There were only 2 parties (which I believe means that I needed 3 copies).
Or the time I sent a single consent order to the court via email and was told I needed to send it X+1 times because there were X parties. When I said they could just print it X+1 times, I was told that wasn't possible.
And woe betide anyone actually trying to phone the court to check something. You can spend literally hours trying to get through.
Closing courts isn't the issue. The issue is the shockingly low quality of some court staff, and an inability to get anything before a judge quickly. It's not possible to offer an efficient service to clients when you're being hamstrung by court errors. The MoJ needs to employ and train more people on the frontline.
"Vara said 97% of citizens will be able to reach their required court within an hour by car, even after the closures" .
This is just like BT Openreach Broadband statistic that they support 98% of the UK population with high speed broadband so should not be split away from the rest of BT. 98% of the population live in highly populated Urban Areas covering around 2% of the UK Land Mass so why care about the last 2% who live in the other 98% of the UK.
As the saying goes, "There are lies, dammed lies and statistics" and it isn't very hard to guess which is which in this case.
No, Anon 4.48, I can't say I do. But it sounds a great idea for the speedy administration of justice. Like all great ideas in this crazy world it will not have lasted any longer than The Green Form Scheme.
Helen, the statement says " Work will either move to an alternative HM Courts & Tribunals Service site, or a decision has been taken to establish suitable alternative local provision before a court will close (These courts are marked by an asterisk)."
When did the present government care about access or justice, let alone the two together!
I think perhaps you mean "asterisk" rather than Asterix (the Gaul) who is a cartoon character.
David, do you remember the days when the courts introduced "Practice Judges" when we could go, without an appointment, sometimes having to queue for a while but at least getting to see a judge and getting our job done?
Unfortunately that system didn't last very long and then went the way of all good ideas.
You actually got through to speak to someone at a County Court?
Were you dreaming per chance?
Bury St Edmunds' Courts appear in the list of those courts not to be closed ( with an interesting asterix next to it) but also are earmarked for closing between July and September of this year.
Could someone who actually knows what is happening explain the significance of the asterix.
Or am I asking too much!
Closing the County Court at Central London and moving it to the RCJ has been about as useful as an ashtray on a motorcycle. We need more efficient Courts, not less that are inefficient. I called Central London yesterday to tell them that a four day trial could be vacated and I was about to submit a Consent Order. I was told I couldn't be put through to the Listings Office as they would only do that if the trial was the next day!!
The poor administration of County Courts drives me to utter distraction.
Getting rid of Courts that are "expensive to maintain"? That explains why the MOJ took on the Manchester Civil Justice Centre at an annual rent of £10m.
"and the maintenance of 'high-quality service provision" - which County Court are they talking about ?
I wholeheartedly agree with those who are expressing their concerns in relation to the time it takes to obtain a court hearing.
My experience is that it is not unusual to be informed that one must wait 6 months to get a 30 minute hearing before a DJ. Surely it must follow that by closing so many courts these waiting periods will increase.
"Vara said 97% of citizens will be able to reach their required court within an hour by car, even after the closures"
Which of course ignores those who are reliant on public transport, which of course includes many of the most vulnerable, and disproportionately hits people involved in Care proceedings and Criminal cases.
So they are closing courts and increasing fees. 1/3 of courts not using their available time (who's fault is that) and why in that case does it take a couple of months to get a 30 minute hearing? Surely that time limit will increase even further when other courts take on further work.
I know what, let's just do away with 1/3 of parliamentary constituencies and MPs. I am sure 400 or so could just as easily work out what is in their own best interests, which is what they seem now to be spending most of their time doing anyway.
Mr Hyde, you may have noticed your home Court at Colchester is on the list to be closed, despite currently being so busy they have a 3 week back log.
In my experience this is because the MOJ don't provide any Judges to hear the cases!
| 2019-04-23T06:39:42 |
https://www.lawgazette.co.uk/news/moj-confirms-86-out-of-91-threatened-courts-will-close/1/5053586.article?PageNo=2
|
0.999214 |
IBM Cognos is a Business Intelligence or BI Suite for enterprise reporting, dashboard, analysis, and scorecard.
IBM Cognos architecture designed for easy to scale, availability and openness.
Following is the image of IBM Cognos Architecture.
1. Web Server: It contains IBM Cognos Web Server contents with one or more IBM Cognos gateway.
2. Application Server: On the application server we can find different types of services. Each service has their own specific role.
Based on the active services application server can handle requests for the report and/or query execution.
The result for any specific service execution forward to end user using the gateway.
3. Database Server: It contains one or more server. The database server is required to store a content store database which IBM Cognos database used to store IBM Cognos components. All the reports, folder, and other information are stored in the content store. Also, data for reporting will reside on database servers.
First, we have a client. The client may be a business user interested in reports, or a developer. All clients used a graphical user interface (GUI) to access IBM Cognos services. IBM Cognos provides two types of user interface. Web-based and Windows-based.
For all the web-based tools client need to use a supporting browser. Using a browser client can access the web-based interface.
Windows-based tools are normally for development and testing purpose. To use a window-based tool, we need to first install and configure it in the local machine. Without proper configuration, we will not able to use it.
All the web and windows tools need to have access to the IBM Cognos gateway. Without proper access to the IBM Cognos gateway, we are not able to use any tools.
All the user action sends to the web server via a gateway. Web Server will process and forward the user's request to an application server for further processing.
An application server processes the user's request and sends it back to the web server. The Web server creates a presentation based on the received response from the application server and displays it to the end user.
We will use the login page. On the login page, we need to provide the user-name and password.
Login page shares this information to the web server.
The Web server encrypts the information, add the necessary environment variables, and send this information to an application server.
The application server forwards the request to authentication services.
Authentication services verify the user and generate a response for authentication.
The application services, send back the authentication response to the web server.
Based on the application server response web server will render a new presentation user interface for the client.
Following are the different services.
| 2019-04-19T10:32:00 |
https://vlemon.com/tutorial/cognos-10-framework-manager/ibm-cognos-architecture
|
0.999998 |
Translation:I do not remember a worse party.
imreza is only a party or also an event?
It can mean an event in general. For example, sport olympics can be referred to as impreza (sportowa).
The suggested solution I read was "I don't remember the worse party." Isn't this a very strange sentence? If you use 'the' it means there's only one (otherwise you'd use 'a'). So if there's only one that's worse, wouldn't it be the worst?
In other words, shouldn't the possible solution "I don't remember the worse party." be removed?
The default sentence has 'a worse party', so 'the' shouldn't be accepted unless you use it yourself. But Duo sometimes does not work how it should. Theoretically you could have mentioned two parties, so one is 'the better one' and the other 'the worse one'.
Having said that, we will think if this isn't indeed too weird to accept.
Nie moge pamientac gorszej imprezy means the same, Correct?
No. Apart from the orthographic mistake in 'pamiętać', it's a literal translation of "I cannot remember". So, effectively, you are saying that you are not (physically?) able to remember a worse party.
Of all the parties I've attended, I can't remember a worse party. You mentioned weird, American English borders on it. We have taken a great language and altered it to the point that it's more slang than sophisticated. I have watched the decline for over sixty years and feel compassion for future generations. Or as our president would say, of course there is always the cellphone. AKA. Making of idiots.
Oh, relax. All languages evolve, have always evolved, and will always evolve. Many things that were once considered formal are now informal, and things that were once informal are now formal. Just because young people are using some new slang - as they always have - is no just cause for panic. The English language will be fine.
"I don't remember the worse party"? I realise I'm being pernickety but couldn't this work? Sounds less natural in english but it definitely is a valid sentence.
A few months ago we've decided that it's just too strange to accept. It's also a course for Polish people learning English, after all.
Then there is the, I can't remember a worse shindig, no end to what you may translate it to in America, I believe the current generation is trying to destroy the language as we were taught it in school seventy years ago. With the advent of the cellphone, the language has has steadily deteriorated into nonsensical mush.
There can be "I cannot remember worse party"!!!
I'm pretty sure you need "a".
There are in fact sentences if translated to English literally sound improper. Though it's not an English lesson, it's important to have good diction. It's all about proper communication regardless of language origin.
I have another question about a word we used back 70 plus years ago. We used "ruszaj" for touch, if still viable, how would it be used. I remember my mom saying, nie ruszaj to.
| 2019-04-23T07:52:26 |
https://forum.duolingo.com/comment/14831756/Nie-pami%C4%99tam-gorszej-imprezy
|
0.999782 |
Review: Cloak -- automatic VPN for privacy on public Wi-Fi - iPhone J.D.
When you use public Wi-Fi, such as the service provided in a hotel or in a coffee shop, it is possible for someone else on the same network to use hacker tools to intercept data going to and from your computer, iPhone or iPad. This makes it possible for a bad guy to get your password as you are logging on to a site, or possibly even to read your confidential attorney-client communications. Is this actually happening to you? Unfortunately, there is no way to know, so all you can do is guard against it. A good solution is to use VPN — a virtual private network — to encrypt the communications from your device to some other location. For example, your law firm may have a way for you to create a VPN to your office, so even if you are on public Wi-Fi the traffic between you and your office is secure, and then you are essentially just as safe using the Internet as you would be in your office. But for this to work, your office must have VPN, and you need to remember to establish the VPN connection.
Cloak is a service that seeks to make it easy and automatic to use a VPN whenever you are using an untrusted public Wi-Fi network. Once you install the app on your iPhone or iPad, whenever you start using a Wi-Fi network, Cloak kicks in automatically and creates a VPN connection to encrypt all of your traffic. You can tell Cloak that a specific Wi-Fi network is trustworthy — such as the one at your home or at your office — and Cloak will remember not to turn on VPN at those locations. But as soon as you connect to Wi-Fi somewhere else, Cloak turns on the VPN to protect you again.
Cloak costs $2.99/month for only 5 GB of data or $9.99/month (or $99.99/year) for unlimited data. There is a free 30 trial if you want to see how the service works, and I've been trying it out for the last few weeks on my iPhone and iPad. (Cloak also works on a Mac, but I haven't tested that.) I'm really impressed with the service. It works behind the scenes so you don't have to worry about it, it doesn't slow down Internet traffic in any way that I could measure, and it provides protection.
When you first start the app, it will switch you over to a part of the Settings app that contains profiles, which is a part of Settings that you may never have used before. Simply click the install button to install the Cloak profile. You will need to enter your device password to confirm the installation.
Now, any time that you connect to a Wi-Fi network that you have not previously told Cloak that you trust, Cloak kicks in and establishes a VPN connection. You probably won't even notice that it happens because it happens quickly, invisibly and in the background. But if you pay attention to the top of your iPhone or iPad screen, you will notice the letters VPN in a box to tell you that you are using a secure VPN connection.
There is a Cloak app on the iPhone and iPad that you can launch to tell Cloak that you trust a network. When you change settings such as adding a new trusted network, Cloak walks you through a simple process to sync your settings so that the same settings are applied to all of your other devices. So if you are using your iPhone and tell your iPhone that you trust your home and your office networks, the next time that you sync settings on your iPad your iPad will also know to trust those networks.
Q: Why should users trust your company?
Q: Couldn't one of your employees access my traffic since I am on your network?
If you click on the links you get very complete answers on why Cloak is trustworthy. The company consists of three guys who used to work at Microsoft — you can see their pictures here — and the whole point of their company is to protect your communications. Obviously, it would destroy their business model if they broke this trust, and they have been doing this since 2012 and have received really good reviews over the years. After reading all of the policies and researching the company, I feel comfortable trusting Cloak to do what it promises to do and protect my privacy — and I certainly trust them much more than I trust all of the random people sitting in the Starbucks with me. Especially that one guy in the corner over there.
Cloak offers one more feature that I haven't found a need to try out, but it might be useful to some of you. When you use a VPN, it can appear to the outside world that you are accessing the Internet from another location. Your iPhone might be in Miami, but a secure tunnel carries your traffic to Seattle so a website might think you are in Seattle. Cloak includes a feature called Transporter that can let your iPhone or iPad appear to be in Australia, Canada, France, Germany, Japan, Netherlands, the United Kingdom or the United States. So if you want to stream video from the BBC that isn't available outside of the U.K., apparently you can make the BBC think that you are in the U.K. Again, not something that interests me, but it is there if it interests you.
I'm impressed enough with this service that when my free 30 day trial ends, I plan to sign up for the basic $2.99/month service to get a sense of whether I use more than 5 GB a month on public Wi-Fi networks. The company also sells through the iPhone app a one week unlimited pass for $3.99, which could be cheaper if the only time that you use public Wi-Fi is when you travel and you don't travel more than twice a month. Just don't forget to buy a pass before you start using the Wi-Fi at the airport, your hotel, etc.
| 2019-04-24T03:59:48 |
https://www.iphonejd.com/iphone_jd/2014/04/review-cloak-vpn.html
|
0.999998 |
Digitization help student to explore and study their academic courses online, as this gives them flexibility and scheduling their learning at their convenience. Kidsfront has prepared unique course material of English Language Verb for UIIC ASSISTANT RECRUITMENT student. This free online English Language study material for UIIC ASSISTANT RECRUITMENT will help students in learning and doing practice on Verb topic of UIIC ASSISTANT RECRUITMENT English Language. The study material on Verb, help UIIC ASSISTANT RECRUITMENT English Language students to learn every aspect of Verb and prepare themselves for exams by doing online test exercise for Verb, as their study progresses in class. Kidsfront provide unique pattern of learning English Language with free online comprehensive study material and loads of UIIC ASSISTANT RECRUITMENT English Language Verb exercise prepared by the highly professionals team. Students can understand Verb concept easily and consolidate their learning by doing practice test on Verb regularly till they excel in English Language Verb.
He said that we are all born to die.
a) He exclaimed, "We were all born to die".
b) He said , "we were all born to die"
c) He said , "we are all born to die"
d) He said, "we have all been born to die".
Correct Answer Is : He said , "we are all born to die"
Solution Is : Direct speech: He said ,"we are all born to die"
He said to me, "I grew up these carrots myself"
Mrs. Shankar said, "I know what it is to be depressed."
"Who now, "they had asked ,will listen to our troubles and protect us from the crocodiles?"
Solution Is : Indirect speech : "They had wanted to know who then would listen to their troubles and protect them from the crocodiles"
A sendence has been given in Active / Passive voice. Out of the four alternatives suggested, select the one whice best expresses the same sentence in Passive / Active voice.
He picked up the telephone and dialled the police number.
a) The telephone was being picked up and the police number dialled.
b) The telephone was picked up and the police number dialled.
c) The telephone had been picked up and the police number dialled.
d) The telephone has been picked up and the police number dialled.
Correct Answer Is : The telephone was picked up and the police number dialled.
A sendence has been given in Direct / Indirect voice. Out of the four alternatives suggested, select the one whice best expresses the same sentence in Indirect / direct voice.
He proposed to his companions that they should not miss the chance to see the last show.
b) He said to his companions, Let us not miss that chance to see the last show.
Correct Answer Is : He said to his companions, Let us not miss that chance to see the last show.
Theteacher said, "Water boils at 100°C."
a) The teacher told water bolied at 100°C.
b) The teacher said that water boiled at 100°C.
c) The teacher said that water boils at 100°C.
d) The teacher told that water boiled at 100°C.
Correct Answer Is : The teacher said that water boils at 100°C.
"Bring me an elephant with one golden tusk," the witch told her minion.
a) The witch requested her minion to bring her an elephant with one golden tusk.
b) The witch told her minion to bring her the elephant with one golden tusk.
c) The witch ordered her minion to bring her an elephant with one golden tusk.
d) The witch advised her minion to bring her an elephant with one golden tusk.
Correct Answer Is : The witch ordered her minion to bring her an elephant with one golden tusk.
"How much will you pay for this cow?" Jack asked the strange man.
a) Jack asked the stranger man how much he would pay for that cow.
b) Jack asked the stranger man how much would he pay for that cow.
c) Jack asked the stranger man how much would he pay for this cow.
d) Jack asked the stranger man how much he will pay for that cow.
Correct Answer Is : Jack asked the stranger man how much he would pay for that cow.
| 2019-04-23T20:55:01 |
http://www.kidsfront.com/govt-jobs-exams/study-material/UIIC+ASSISTANT+RECRUITMENT-+English+Language-Verb-p0.html
|
0.999983 |
Subtitle: "If people seek the 'tune' (the music's hidden message) then Lucifer will lead them to Absolute Truth." An entire nation has been seduced by these songs, especially The Stairway to Heaven. "Back to the old gods", this Rock Music urges. Written by Tom Friend, author of the book, "Fallen Angel".
The story of this book is very complex; it is a detailed effort to present to the reader the facts that Jimmy Page and Led Zeppelin boldly propagated the teachings of Black Magician Aleister Crowley and of the satanic, magical cult he founded, the “Argenteum Astrum” -- which means “Silver Star" . The major doctrines of Aleister Crowley’s Silver Star are presented in the most famous and influential songs in the Led Zeppelin catalog -- including Stairway to Heaven. This rock song has repeatedly been voted the most popular song in the history of rock and roll, on both sides of the Atlantic Ocean. To this day, the album with Stairway to Heaven on it sells four to five thousand copies a week in the United States alone. This album has sold a total of over 23 million copies in the United States; second only to Michael Jackson’s “Thriller.” The Stairway to Heaven album is loaded with Aleister Crowley’s bold Luciferian teachings.
The worship of Lucifer, in the form of Pan, was a theme through all of Aleister Crowley’s writings. Virtually all of Rock-n-Roll music recognizes Aleister Crowley as the spiritual headwaters from which the lyrics and the antics flow.
Edward Alexander “Aleister” Crowley was born in 1875 in England. He was born into a family that was part of a strict sect of evangelical Christians called “The Plymouth Brethren.” Crowley’s father was a prominent member of this sect. He was a preacher who used to travel to evangelize. The Crowley family was also wealthy because Aleister's Dad was also a very successful brewer of alcohol. One can immediately see the obvious, apparent contradiction in being an evangelist of the Gospel and a brewer of alcohol. This was a contradiction that little Aleister didn’t fail to recognize. Nevertheless, Crowley was encouraged by his father to love God and to want to grow in the Christian faith. Crowley was receptive to his father’s guidance initially. He used to love reading the Bible as a boy. He did state in his “Confessions” that he was fascinated by presentations of Satan and the Anti-Christ in the Book of Revelation. Nevertheless, he continued to study the Bible.
However, his father passed away when he was eleven years old. His family had placed him in a boarding school run by some members of "The Plymouth Brethren". According to Aleister, this school was supposed to be run by Christians -- they had constant chapel services, and a Reverend. However, Crowley documented that the school was dominated with liars, homosexuals, and a number of other forms of wickedness explained in detail in Fallen Angel. This terrible hypocrisy was enough to turn Aleister Crowley against Christianity for life. Even before he attended this school, he admitted to possessing a love for “the enemies of Heaven” in the Book of Revelation; the events at the boarding school gave him the spiritual "green light" he sought to give himself over to Satan.
Aleister Crowley went on to write dozens of books on the occult throughout his life, promoting Black Magic and all forms of spiritual wickedness, including astral projection, divination, rituals, human sacrifice, the Universality of Magick, and practically anything else that would help him achieve his stated goal -- the destruction of the Christian faith.
The statement that Christ has a phony crown reflects their belief that Lucifer is god, as taught them by Aleister Crowley. Another thing that Crowley emphasized to his readers was that musicians in Magick would be able to be 'musical mediums'. When speaking of “Stairway to Heaven", Robert Plant claimed to be a 'musical medium'. He said the following about how he received the lyrics to the song: Plant recalled.
The word, "pipes" (Strong's #5345) clearly refers to music and to the flute, as we see when we examine the context of this same word in other Old Testament Scriptures.
From these verses, you can easily see that God is stating that He created Lucifer as a musical being in Ezekiel 28:13. In much occult lore, the Greek god, Pan, is this musical representation of Lucifer, the god of "pipes" -- the Piper.
The worship of Lucifer, in the form of Pan, was a theme through all of Aleister Crowley’s writings. Jimmy Page has always been an open student of Crowley the Black Magician, and Led Zeppelin was his band. Following the teachings of Crowley, Jimmy Page and company were successful in placing his doctrines in many of their songs, including “Ramble On,” “Black Dog,” “The Battle of Evermore,” “Stairway to Heaven,” “No Quarter,” “The Song Remains the Same,” “Kashmir,” “In The Light,” “Achilles Last Stand,” and “Nobody’s Fault But Mine,” just to name a few.
In Fallen Angel, you will realize the entire spectrum of occult teachings presented by Aleister Crowley; when combined with other renowned occult authors, you will understand the truth about Led Zeppelin’s worship, and propagation, of Lucifer. These occult teachings are thoroughly refuted by this author, presenting the truth of the gospel, including the Biblical Absolutes of Truth. This book has been written to inevitably glorify Lord Jesus Christ, especially in the heart and mind of the rock fan who needs to receive Christ as Lord and Savior. The evidence is overwhelming, yet factual, with the Word of God presented as the absolute authority for faith and practice. The objective reader, if a Christian, will be thoroughly equipped to witness to the rock culture; and the lost sinner will see and appreciate his need for salvation in our Lord Jesus Christ.
The book demonstrates that the End of the Age is clearly upon us, and that Rock Music is playing the "Pied Piper" to hundreds of millions of people, leading them into a worship of Lucifer. Are you spiritually ready? Is your family? Are you adequately protecting your loved ones? This is the reason for this ministry, to enable you to first understand the peril facing you, and then help you develop strategies to warn and protect your loved ones. Once you have been thoroughly trained, you can also use your knowledge as a means to open the door of discussion with an unsaved person. I have been able to use it many times, and have seen people come to Jesus Christ as a result. These perilous times are also a time when we can reach many souls for Jesus Christ, making an eternal difference.
| 2019-04-26T07:41:28 |
http://www.cuttingedge.org/news/n1957.cfm
|
0.998997 |
PRINCETON, NJ -- Americans believe that the U.S. Army and Marines are the most important to national defense, followed by the Air Force and Navy. This differs from the years prior to the start of the Iraq war, when Americans named the Air Force as the most important branch of the armed forces.
Gallup first asked Americans about the importance of the military branches in the 1940s, and began using the current question wording in 2001. Regardless of exact question format, at all points prior to 2004, Americans -- by very substantial margins -- have named the Air Force as the most important branch of the military.
A major shift occurred between 2002 and 2004, concomitant with the beginning of the Iraq war in which the primary focus of news coverage was on the actions of U.S. ground forces. The percentage of Americans naming the Air Force as most important military branch declined in 2004, while Americans placed more importance on the Army and Marines.
This change continues, with the Army and Marines essentially tied in Gallup's June 9-12 survey as most important to national defense, while the perceived importance of the Air Force has dropped further, from 23% in 2004 to 17% today. The percentage of Americans naming the Navy as most important has dropped from 17% in 2002 to 11% today. Gallup first included the Coast Guard in the question wording in 2002, but this branch has received few mentions as most important since that point.
Gallup first measured Americans' perceptions of the importance of the branches of the armed forces in June 1949, using this question wording: "If the United States should get into another World War, which branch of the Armed Forces do you think would play the most important part in winning the war -- the Army, the Navy, or the Air Force?"
At that time, 81% of Americans chose the Air Force, followed by 6% who named the Army and 4% the Navy. In August 1951, in the middle of the Korean War, the Air Force continued to dominate these "most important" perceptions, with 70% choosing it. And, in 1960, before an election in which two former naval officers -- John F. Kennedy and Richard Nixon -- were battling for the presidency, 62% of Americans named the Air Force as most important, while 5% picked the Army and 6% the Navy.
Americans have named the Marines as the most prestigious branch of the armed forces in each of four surveys conducted between 2001 and 2011. Thirty-six percent named the Marines as most prestigious in 2001, while 46% do today.
Americans' view of the Air Force's prestige has dropped over this time period. In 2001 the Air Force was a few percentage points behind the Marines as most prestigious. Now, in 2011, the Army has moved into second place in prestige behind the Marines, while the percentage mentioning the Air Force has dropped by more than half, from 32% to 15%.
For many decades, stretching from 1949 to 2002, Air Force generals were no doubt pleased to find that the average American viewed the Air Force as the most important branch of the service -- by very significant margins. This occurred despite the predominant role of ground forces in the Vietnam War and the successful effort led by U.S. ground forces to push Saddam Hussein's Iraqi army out of Kuwait in early 1991.
Now, in more recent years, it is Army and Marine generals' turn to be pleased, as Americans view these two branches of the military as the most important, with the Air Force and the Navy lagging behind.
The shift in perceptions occurred between 2002 and 2003, when the U.S. first became involved in the Iraq war. The Air Force and the Navy have been heavily involved in fighting these two wars, of course, but despite this, public perceptions of the importance of these two service branches have waned. The high levels of visibility given to the role of the Navy SEALs in the death of terrorist leader Osama bin Laden in May, additionally, apparently has done little to significantly enhance the perceived importance or prestige of the Navy.
The results reported in this analysis for the perceived importance and perceived prestige of military branches are each based on random half samples of 496 and 524 national adults, respectively. For each of these samples, one can say with 95% confidence that the maximum margin of sampling error is ±5 percentage points.
Nearly all Americans approve of the U.S. military action that resulted in the death of Osama bin Laden on Sunday, and 79% say the action was "extremely" or "very important" to the U.S. Americans credit the U.S. military and the CIA most, followed by President Barack Obama and then former President George W. Bush.
Americans give their best assessment since July 2009 about how things are going for the U.S. in Afghanistan, with 47% saying things are going well and 49% saying things are going badly. A majority of Americans continue to say the U.S. did not make a mistake in sending military forces to Afghanistan.
Americans are more likely to approve (47%) than disapprove (37%) of U.S. military action against Libya. That level of support is low compared with the strong initial backing Americans have given to many other recent U.S. military actions.
| 2019-04-25T01:55:48 |
https://news.gallup.com/poll/148127/Americans-Army-Marines-Important-Defense.aspx?utm_source=tagrss&utm_medium=rss&utm_campaign=syndication
|
0.999973 |
The input is a positive integer C and n objects whose sizes are positive integers s1, s2,..., sn.
Optimization Problem: Among all subsets of objects with sum at most C, what is the largest subset sum?
Decision Problem: Is there a subset of objects whose sizes add up to exactly C?
| 2019-04-21T02:56:23 |
http://lcm.csa.iisc.ernet.in/dsa/node221.html
|
0.9985 |
The article below appears in today's San Jose Mercury News (Sunday, Dec. 3, 2006.) It was written by columnist Scott Herhold.
In the primitive days of politics, neighbors trying to fight an environmentally questionable project on 1,000 acres near their homes might have to depend on word of mouth, a telephone chain to summon the faithful to endless public meetings.
But this is the era of iPods, Blackberries and Google Earth, when the digital can trump the verbal, when even a national leader can be recruited to a local cause through the power of a display on the Web.
That brings me to a tale that includes former Vice President Al Gore, the trees above Lexington Reservoir and the ingenuity of Rebecca Moore, a Google Earth engineer who lives in the Los Gatos mountains.
Here's the background: Citing the need for fire protection, the San Jose Water Co. has submitted a plan to log 1,000 acres of its land in the mountains above Los Gatos.
The plan has prompted sustained protest from the neighbors, who insist that the water company's plans will degrade the water quality and increase the danger of fire.
I don't pretend to understand the details, which is where the good and bad of logging resides. From my quick reading, it strikes me that the neighbors have raised legitimate questions about plans to cut larger trees, defined as more than 24 inches in diameter.
But this story isn't so much about the rights and wrongs of the issue as it is about the way the debate has unfolded politically and technologically. The two strands merge.
About 15 months ago, in September 2005, the water company sent a notice to neighbors of their intent to harvest timber, complete with a somewhat sketchy map. One of the people who got it was Rebecca Moore, who worked on developing Google Earth, the cool Web feature that lets you zoom down from outer space to see your back yard.
Other folks might have shrugged. Moore created a virtual Google Earth ``flyover,'' which gives you the impression of traversing the logging area in a helicopter at 1,000 feet. You can see it yourself by going to www.mountainresource.org/nail, and scrolling down until you see the Google Earth image.
Perhaps her best-known convert came to the cause eight months later. Gore, a senior adviser to Google and the central figure in the movie ``An Inconvenient Truth,'' wanted to see how Google Earth worked. After showing him a couple of global examples, Moore showed him the Los Gatos Creek flyover.
In response, the Big Creek Lumber Co., which is planning the cutting on San Jose Water's property, has said it will formally invite Gore to come look at other logging it has done.
``We think what Big Creek is doing is precisely the kind of land stewardship and responsible forward-looking business practices that he's been advocating for years,'' said Bob Berlage, a spokesman for Big Creek, a family-run firm that has a reputation for not clear-cutting forests.
So the kindling's in the fire. Or the logs are piled high. Whatever. With the support of the inventor of the Internet at stake, could it be anything other than a flaming controversy?
| 2019-04-25T20:52:30 |
http://www.mountainresource.org/nail/blog/technology-builds-bigger-soapbox-sj-mercury-news
|
0.999991 |
Query: Should I sell my spare Amiga 1200?
I am in New Jersey and have two NTSC Amiga 1200s, neither has been recapped and I have an irrational fear that as soon as I sell one the other will implode.
I might use the proceeds from sale to recap the other Amiga 1200.
Is it a reasonable DIY job, or should I have it sent off to be recapped?
I’ve recapped a few, and while not particularly difficult, you do need to be careful as traces can easily lift if you apply too much heat or apply for too long. Contamination makes the solder more difficult to manipulate leading to the overapplication of heat and time, so you need to eyeball the board to make a sensible determination.
Unless you’re comfortable repairing potentially lifted traces, you might have someone else do the work.
EzdineG wrote: I’ve recapped a few, and while not particularly difficult, you do need to be careful as traces can easily lift if you apply too much heat or apply for too long. Contamination makes the solder more difficult to manipulate leading to the overapplication of heat and time, so you need to eyeball the board to make a sensible determination.
Thanks for the advice. Yeah I am hardly an expert at soldering, if that is the case I am leaning to send it away to someone who's done it before.
I have recapped both of my 1200's, an NTSC and a PAL. If you have SMD cap experience it should be no problem. The toughest areas are the caps behind the keyboard connector and in that area, as well as a couple of the through-hole caps that are on a ground plane. They require extra heat and even better a little fresh solder added in before trying to desolder.
That said, If you don't have experience with SMD caps, the A1200 is not the one you want to be your first one.
What kind of condition are they in? All stock?
One is a Commodore era A1200 with a flash drive installed internally (otherwise stock), but with a broken enter key. The other is an Escom era A1200 also with a flash drive installed, as well as a 25mhz 030 and 4mb Ram expansion. Both have working floppies.
The Escom is more yellowed, but I’m leaning toward selling the Commodore era A1200 unless there are any downsides to the later units.
I have owned one of each of those models in the past two years. Both of mine were mint (the ESCOM model was NIB when I got it - I was the first person to ever open it).
That being said, I liked the look of the original C= 1200, and it had an original (tiny) mechanical hard drive that still worked. Thing was cool. The ESCOM model I had was a bright white case. Anyway, they were both good machines and my preference doesn't have any particular logic behind it just personal preference.
Curious - did you ever invest in the A1200 keycap Kickstarter that is on the verge of shipping?
intric8 wrote: I have owned one of each of those models in the past two years. Both of mine were mint (the ESCOM model was NIB when I got it - I was the first person to ever open it).
I did not. I have a problem with the A1200, although I like the idea of AGA and occasionally play AGA games, I always go back to my A500 running WB 1.3 because it is more nostalgic for me.
So I have resisted the urge to invest a lot of money into my A1200s, whereas at least one of my A500s may get a new case and keyboard at some point.
I have a problem with the A1200, although I like the idea of AGA and occasionally play AGA games, I always go back to my A500 running WB 1.3 because it is more nostalgic for me.
I mean, that's what matters.
For me, I could sneeze and AGA would vanish and I'd never even notice. My SNES blows it away from a game machine perspective. WB 1.3 does virtually everything I need it to do, and it does it very well - whether it be games, software applications, hardware needs - everything. There are some nice-to-haves with later OSes and chipsets, but not enough to get me to go there as those whistles and bells usually mean compromising compatibility. I've nothing against it, either, I just prefer the older machines and experiences. So I totally hear ya. And I know several folks who grew up on the 1200 as their first Amiga and understand their perspective, too. There's no singular "right path" - just yours.
So I bought into the keycap program (2 years ago? Man, it's been a while) and don't even own a A1200 anymore.
The fullset is compatible with ALL Amiga classic keyboards.
EXCEPT for the A1000 and some early versions of A2000 or A500 not using Mitsumi Hybrid Switches.
I mean, that's probably going to kick me out even though I bought a set. I only use the A1000 and A2000 today. I do have 2 A2000 keyboards, but I've no idea what kind of switches they use. Maybe they'll work after all I but won't really know until the keycaps finally arrive.
If I had only one all-in-one to choose, though, it would be the 500 for me. It just works with the software I want to use.
I didn't realize you got rid of your A1200. Or maybe I did and just forgot. I have been contemplating getting rid of my PAL 1200. I really don't need 2. I bought the PAL one until I found an NTSC machine as that is what I had back then. It doesn't really matter, though, because you can easily boot into PAL mode or NTSC mode from either. But my NTSC machine is not yellowed in the least bit.
I completely agree. The A500 was the first Amiga I owned. I had more fun on that than any other Amiga. Everything was new and a learning experience. BBS, Graphics, Demos, etc, all experienced on A500 first. When I bought the A1200 on release, I think it was at least a year until I had software that actually used the AGA. I think IIRC, it was an update to Vista Pro.
| 2019-04-25T18:04:13 |
https://amigalove.com/viewtopic.php?f=10&t=903&p=5300
|
0.998856 |
2. What feature is commonly seen in the sequences recognized by type II restriction enzymes?
The recognition sequences are palindromic, and 4−8 base pairs long.
3. What normal role do restriction enzymes play in bacteria? How do bacteria protect their own DNA from the action of restriction enzymes?
Restriction enzymes cut foreign DNA, such as viral DNA, into fragments. Bacteria protect their own DNA by modifying bases, usually by methylation, at the recognition sites.
4. Explain how gel electrophoresis is used to separate DNA fragments of different lengths.
Gel electrophoresis uses an electric field to drive DNA molecules through a gel that acts as a molecular sieve. The gel is an aqueous matrix of agarose or polyacrylamide. DNA molecules are loaded into a slot or well at one end of the gel. When an electric field is applied, the negatively charged DNA molecules migrate toward the positive electrode. Shorter DNA molecules are less hindered by the agarose or polyacrylamide matrix and migrate faster than longer DNA molecules, which must wind their way around obstacles and through the pores in the gel matrix.
5. After DNA fragments have been separated by gel electrophoresis, how can they be visualized?
DNA molecules can be visualized by staining with a fluorescent dye, such as ethidium bromide, that intercalates between the stacked bases of the DNA double helix, and the dye-DNA complex fluoresces when irradiated with an ultraviolet light source. Alternatively, they can be visualized by attaching radioactive or chemical labels to the DNA before it is placed in the gel.
6. What is the purpose of Southern blotting? How is it carried out?
Southern blotting is used to detect and visualize specific DNA fragments that have a sequence complementary to a labeled DNA probe. DNA is first cleaved into fragments with restriction endonucleases. The fragments are separated by size via gel electrophoresis. These fragments are then denatured and transferred by blotting onto the surface of a membrane filter. The membrane filter now has single-stranded DNA fragments bound to its surface, separated by size as in the gel. The filter is then incubated with a solution containing a denatured, labeled probe DNA. The probe DNA hybridizes to its complementary DNA on the filter. After washing away excess unbound probes, the labeled probe hybridized to the DNA on the filter can be detected using the appropriate methods to visualize the label. For radioactively labeled probes, the bound probe is detected by exposure to X-ray film. Other probe labeling methods detect bound probe using enzymatic reactions that generate luminescence or color.
7. Give three important characteristics of cloning vectors.
9. Briefly explain how an antibiotic-resistance gene and the lacZ gene can be used to determine which cells contain a particular plasmid.
Many plasmids designed as cloning vectors carry a gene for antibiotic resistance and the lacZ gene. The lacZ gene on the plasmid has been engineered to contain multiple unique restriction sites. Foreign DNAs are inserted into one of the unique restriction sites in the lacZ gene of plasmids and the plasmids are transformed into E. coli cells lacking a functional lacZ gene. Transformed cells are plated on a medium containing the appropriate antibiotic to select for cells that carry the plasmid. The medium also contains an inducer for the lac operon, so the cells express the lacZ gene, and X-gal, a substrate for beta-galactosidase that will turn blue when cleaved by β-galactosidase. The colonies that carry plasmid without foreign DNA inserts will have intact lacZ genes, make functional β-galactosidase, cleave X-gal, and turn blue. Colonies that carry plasmid with foreign DNA inserts will not make functional β-galactosidase because the lacZ gene is disrupted by the foreign DNA insert. They will remain white. Thus, cells carrying plasmids with inserts will form white colonies. This is known as blue/white selection.
10. Briefly explain how the polymerase chain reaction is used to amplify a specific DNA sequence.
First, the double-stranded template DNA is denatured by high temperature. Then, primers corresponding to the ends of the DNA sequence to be amplified are annealed to the single-stranded DNA template strands. These primers are extended by a thermostable DNA polymerase so that the target DNA sequence is duplicated. These steps are repeated 30 times or more. Each cycle of denaturation, primer annealing, and extension results in doubling the number of copies of the target sequence between the primers.
12. How does a genomic library differ from a cDNA library?
A genomic library is generated by cloning fragments of chromosomal DNA into a cloning vector. Chromosomal DNA is randomly fragmented by shearing or by partial digestion with a restriction enzyme. A cDNA library is made from mRNA sequences. Cellular mRNAs are isolated and then reverse transcriptase is used to copy the mRNA sequences to cDNA, which are cloned into plasmid or phage vectors. Therefore a cDNA library only represents the genes expressed in the tissues of origin for the RNA.
13. How are probes used to screen DNA libraries?
The DNA library must first be plated out, either as colonies for plasmid libraries or phage plaques on a bacterial lawn for phage libraries. The colonies or plaques are transferred to membrane filters. A nucleic acid probe can be used to identify colonies or phage plaques that contain identical or similar sequences by hybridization.
14. Briefly explain in situ hybridization, giving some applications of this technique.
In situ hybridization involves hybridization of radiolabeled or fluorescently labeled DNA or RNA probes to DNA or RNA molecules that are still in the cell. This technique can be used to visualize the expression of specific mRNAs in different cells and tissues and the location of genes on metaphase or polytene chromosomes.
17. What is the purpose of the dideoxynucleoside triphosphate in the dideoxy sequencing reaction?
Dideoxynucleoside triphosphates (ddNTPs) act as a substrate for DNA polymerase but cause termination of DNA synthesis when they are incorporated, because they lack the 3′-OH for the addition of the next nucleotide. Mixed with regular dNTPs, fluorescently labeled ddNTPs generate a series of DNA fragments that have terminated at every nucleotide position along the template DNA molecule being sequenced. These fragments can be separated by gel electrophoresis. Because each of the four ddNTPs carries a different fluorescent label, a laser detector scanning near the end of the gel can distinguish which base terminates each fragment. Reading the fragments from shorter to longer, an automated DNA sequencer can determine the sequence of the template DNA molecule.
18. What is DNA fingerprinting? What types of sequences are examined in DNA fingerprinting?
DNA fingerprinting is the typing of an individual for genetic markers at highly variable loci. This is useful for forensic investigations, to determine whether the suspect could have contributed to the evidentiary DNA obtained from blood or other bodily fluids found at the scene of a crime. Other applications include paternity testing and the identification of bodily remains.
The first loci used for DNA fingerprinting were variable number of tandem repeat (VNTR) loci; these consist of short tandem repeat sequences located in introns or spacer regions between genes. The number of repeat sequences at the locus does not affect the phenotype of the individual in any discernible way, so the numbers of repeats at each of these loci are highly variable in the population. More recently, loci with smaller repeat sequences of just a few nucleotides, called short tandem repeats (STRs), have been adopted because they can be amplified by PCR. The variable number of repeats creates PCR fragments of different sizes. Genotyping at 13 to 15 of these unlinked STR loci can identify one individual among trillions of potential genotypes.
19. How does a reverse genetics approach differ from a forward genetics approach?
Forward genetics begins with mutant phenotype and proceeds toward cloning and characterization of the DNA encoding the gene. Reverse genetics begins with the DNA sequence of a gene, then generates specific mutations within that gene and requires the reinsertion of the modified gene back into cells or the whole organism to characterize the functions of the gene through new mutant phenotypes.
20. Briefly explain how site-directed mutagenesis is carried out.
In oligonucleotide-directed mutagenesis, an oligonucleotide containing the desired mutation in the sequence is synthesized. This mutant oligonucleotide is annealed to denatured target DNA template and used to direct DNA synthesis. The result is a double-stranded DNA molecule with a mismatch at the site to be mutated. When transformed into bacterial cells, bacterial repair enzymes will convert the molecule to the mutant form about 50% of the time.
21. What are knockout mice, how are they produced, and for what are they used?
Knockout mice have a target gene disrupted or deleted ("knocked out"). First, the target gene is cloned. The middle portion of the gene is replaced with a selectable marker, typically the neo gene that confers resistance to the drug G418. This construct is then introduced back into mouse embryonic stem cells and cells with G418 resistance are selected. The surviving cells are screened for cells where the chromosomal copy of the target gene has been replaced with the neo-containing construct by homologous recombination of the flanking sequences. These embryonic stem cells are then injected into mouse blastocyst-stage embryos and these chimeric embryos are transferred to the uterus of a pseudopregnant female mouse. The knockout cells will participate in the formation of many tissues in the mouse fetus, including germ-line cells. The chimeric offspring are interbred to produce offspring that are homozygous for the knockout allele.
The phenotypes of the knockout mice provide information about the function of the gene.
22. How is RNA interference used in the analysis of gene function?
RNA interference is one potential reverse genetics approach to analyze gene function, by specifically repressing expression of that gene. Double-stranded RNA may be injected directly into a cell or organism or the cell or organism may be genetically modified to express a double-stranded RNA molecule corresponding to the target gene.
23. What is gene therapy?
Gene therapy is the correction of a defective gene by either gene replacement or the addition of a wild-type copy of the gene. For this to work, enough of the cells of the critically affected tissues or organs must be transformed with the functional copy of the gene to restore normal physiology.
26. How often, on average, would you expect a type II restriction endonuclease to cut a DNA molecule if the recognition sequence for the enzyme had 5 bp? (Assume that the four types of bases are equally likely to be found in the DNA and that the bases in a recognition sequence are independent.) How often would the endonuclease cut the DNA if the recognition sequence had 8 bp?
Because DNA has four different bases, the frequency of any sequence of n bases is equal to 1/(4n). A five-bp recognition sequence will occur with a frequency of 1/(45), or once every 1024 bp. An 8-bp recognition sequence will occur with a frequency of 1 per 48, or 65,536 bp.
28. Will restriction sites for an enzyme that has 4 bp in its restriction site be closer together, farther apart, or similarly spaced, on average, compared with those of an enzyme that has 6 bp in its restriction site? Explain your reasoning.
The restriction sites for an enzyme with a 4-bp recognition sequence should be spaced closer together than the sites for an enzyme with a 6-bp recognition sequence. The 4- bp recognition sequence will occur with an average frequency of once every 44 = 256 bp, whereas the 6-bp recognition sequence will occur with an average frequency of once every 46 = 4096 bp.
*29. About 60% of the base pairs in a human DNA molecule are AT. If the human genome has 3.2 billion base pairs of DNA, about how many times will the following restriction sites be present?
We must first calculate the frequency of each base. Given that AT base pairs consist 60% of the DNA, we deduce that the frequency of A is 0.3 and frequency of T is 0.3. The GC base pairs must consist of 40% of the DNA; therefore, the frequency of G is 0.2 and the frequency of C is 0.2.
32. A geneticist uses a plasmid for cloning that has the lacZ gene and a gene that confers resistance to penicillin. The geneticist inserts a piece of foreign DNA into a restriction site that is located within the lacZ gene and uses the plasmid to transform bacteria. Explain how the geneticist can identify bacteria that contain a copy of a plasmid with the foreign DNA.
The geneticist should plate the bacteria on agar medium containing penicillin to select for cells that have taken up the plasmid. The medium should also have X-gal and an inducer of the lac operon, such as IPTG or even lactose. Cells that have taken up a plasmid without foreign DNA will have an intact lacZ gene, produce functional β- galactosidase, and cleave X-gal to make a blue dye. These colonies will turn blue. In contrast, cells that have taken up a plasmid containing a foreign DNA inserted into the lacZ gene will be unable to make functional β-galactosidase. These colonies will be white.
*35. A molecular biologist wants to isolate a gene from a scorpion that encodes the deadly toxin found in its stinger, with the ultimate purpose of transferring this gene to bacteria and producing the toxin for use as a commercial pesticide. Isolating the gene requires a DNA library. Should the genetic engineer create a genomic library or a cDNA library? Explain your reasoning.
A cDNA library, created from mRNA isolated from the venom gland. Bacteria cannot splice introns. If the engineer wants to express the toxin in bacteria, then he needs a cDNA sequence that has been reverse transcribed from mRNA, and therefore has no intron sequences. The venom gland must be the source of the mRNA for cDNA synthesis, so that the cDNA library will be enriched for toxin cDNAs.
How many different probes must be synthesized to be certain that you will find the correct cDNA sequence that specifies the protein?
A probe of 18 nucleotides must be based on six amino acids. The six amino acid stretch with the least degeneracy is Val-Tyr-Lys-Ala-Lys-Trp. This sequence avoids the amino acids arg and leu, which have six codons each.
Val and Ala have four codons each, Tyr and Lys have two codons each, and Trp has one codon. Therefore, there are 4 × 2 × 2 × 4 × 2 × 1 possible sequences, or 128.
37. In Figure 19.22 Bob and Joe are each homozygous for different restriction fragment patterns. How many bands would you expect to see on the gel if a person was heterozygous for the A and B patterns? Explain your reasoning.
A person heterozygous for the A and B patterns would exhibit four bands: two bands associated with pattern B, and the two additional bands represented by pattern A. This is because a heterozygous person would have alleles for both the A and B patterns. The DNA for the A pattern would be cut into three fragments. The DNA for the B pattern would be cut into two fragments, one of which was the same as that for the A pattern and one which was different. Thus, the heterozygote would have DNA cut into four fragments and would exhibit four different bands.
42. You have discovered a gene in mice that is similar to a gene in yeast. How might you determine whether this gene is essential for development in mice?
This gene must first be cloned, possibly by using the yeast gene as a probe to screen a mouse genomic DNA library. The cloned gene is then used to create a knockout mouse strain. A neo gene is inserted into a cloned exon sequence, followed by the tk (thymidine kinase) gene. This construct is then introduced into mouse embryonic stem cells. Selection of G418 resistant and gancyclovir resistant cells will identify transformants where homologous recombination resulted in insertion of the neo gene into the target mouse gene exon. These knockout stem cells are injected into blastocyst stage embryos, which are then transferred to the uterus of a pseudopregnant mouse. The progeny are tested for the presence of the knockout allele. Progeny with the knockout allele are interbred. If the gene is essential for embryonic development, no homozygous knockout mice will be born. The arrested or spontaneously aborted fetuses can then be examined to determine how development has gone awry in fetuses that are homozygous for the knockout allele.
| 2019-04-26T00:16:51 |
https://quizlet.com/107037455/chapter-19-solutions-manual-flash-cards/
|
0.999999 |
Can I contribute a photo?
Frederick Holbrook was born c. 1864 in Belper, Derbyshire. By the time of the 1881 Census he was living with his older brother Samuel at Common Side, Belper, and working as an engine man. Trade directories later show Frederick Holbrook with studios at Albert Street (1895) and George Street (1912) in Belper. The 1901 Census indicates that he was working as a photographer on his own account, and boarding with a family named Street in Chapel Street, Belper. Adamson (1997) gives dates of 1906-1912 for the operation of the George Street premises. A postcard photo taken c. 1915 suggests he also had a studio at New Road, and a 1925 directory gives the address of a studio at 14 Field Row in Belper, a building which, according to the Belper Research web site, burnt down.
No further information about this photo.
Notes by CM: This is Fanny Whiteman, my grandfather Amos Whiteman's sister from Milford. The family lived at Hopping Hill in her early days.
Annie Rushworth - baptized 15 October 1856 at Heslington, Yorkshire, married Alfred Watherston from Heslington, Yorkshire (I've yet to find a marriage record). I don't when she died but she had moved to Blackpool, England in about 1924 and I have mail that was sent to her in 1938. My mother (b.1933) remembers an old woman living upstairs to his gran's who use to bang on the floor with a cane when they made too much noise.
Alathea Mary Yates Watherston - born 20 December 1886 at Leeds, Yorkshire, married Thomas Alton 20 April 1908 at Heage, Derbyshire, died 21 March 1966 at Blackpool, Lancashire.
Freda Mary Alton - born 1909 at Heage, Derbyshire, never married, died 27 April 1991 at Blackpool, Lancashire.
Blackpool, Lancashire, died 14 June 1965 at Blackpool, Lancashire.
My great-grandfather Thomas Alton, Alathea's husband, died of pneumonia at the age of 34, on 1 February 1917, after six weeks of bootcamp. I believe this photo was taken for him to keep with him if he was sent to the trenches. All the photo's etc., are thanks to Freda. They were unknown until discovered in a chest in 1987 when she was being moved into a nursing home.
Anon (1895) Kelly's Directory of Derbyshire, Kelly & Co. Limited, London.
Anon (1912) Kelly's Directory of Derbyshire 1912, Kelly's Directories Ltd., London.
| 2019-04-20T20:20:44 |
http://freepages.rootsweb.com/~brett/genealogy/photos/fholbrook.html
|
0.999833 |
I am trying to work out how I can achieve a MAX() type selection on a range of cells, but limit the cells over which I perform the MAX() according to the value of the adjacent cell.
Cells A1:A9 contain a random mix of "Apples", "Oranges" and "Bananas". Each cell in B1:B9 contain a number of fruit of the type contained in the adjacent "A" cell.
What I want is to put in Cell D1 (for example) the MAX() of the values in B1:B9, but only for (say) "Apples".
If I do MAX(B1:B9), I (rightly) get the simple highest value, regardless fo the type of fruit. What I need is some way to filter on the fruit and MAX() only matching cells.
Is there a reasonably simple way to do this without modifying (e.g. by sorting) the source data?
I see how this should work as I can find references in Excel fora about this funtion, but it isn't in my Libre Calc.
and executed it as an array formula by pressing Ctrl-Shift-Enter.
Not sure if this is the best way, but it got me what I want. I'm a happy boy now.
I don't think you need to enter this as an array function, although it may depend on the version of LibreOffice. The IF function in this case returns an array, but then MAX takes the array and returns a single value.
If I enter the formula as above and just press "Enter", it fails with "#VALUE!". If I press "Ctrl+Shift+Enter" it works. I'm using Libre Calc Version: 5.1.5.2 under Windows 7.
I'm using LibreOffice 5.2.1.2 x64 on Windows 8.1. I think I needed Ctrl+Shift+Enter on some earlier version of LO but I forget which one.
I downloaded your spreadsheet, but it errors against cell E1, which contains"=com.microsoft.maxifs(B1:B9, A1:A9, D1)".
I assume the "com.microsoft.maxifs" is some external funtion (probably from excel?), but I don't have this on my machine. I see how it works, but I need to know how to get this funtion into Libre Calc.
| 2019-04-22T18:59:32 |
https://ask.libreoffice.org/en/question/78528/how-to-get-highest-value-using-libre-calc/?answer=78531
|
0.999929 |
Study the essential vocabulary and translate the illustrative examples into Russian.
A. Insert the correct preposition. Compare it with the one we use in Russian.
the length of days, but in the use we make of them; a man may live long, yet get little from life (M. Montaigne). 43. Though he prided himself on trusting no one, he always accepted at face value any friendly gesture that was offered to him. 44. While the New Yorker can appreciate the beauties of nature where he can forget the urgent problems of the day, he seems to be unaffected by the joys of country life. 45. The expeditionNvas in urgent need of supplies. 46. ??Everything urgent had been dealt with by her efficient secretary. 47. Old Jolyon could hardly resist June's urgent requests. 48. "Well, a good novel is real, far more significant than most of the highbrow stuff - so-called", he said, taking a little time to answer. 49. "He knows his stuff ', said Monsier Poirot with evident approval. 50. My father was a stuffy man. He always wore dark suits and ugly ties, and was for ever pursing his lips and wrinkling up his forehead before he said anything . 51. He stuffed his ears with cotton wool not to be distracted by the noise.
солодощі; зелень; фарширована риба; напихати дитини; запихати речі в чемодан; пхати що-л. в кишеню.
1. I could never understand what caused their quarrel. 2. Very often an emergency reveals a person's main qualities. 3. The noise of the gun scared hundreds of birds. 4. She is very difficult to please, always complaining or worrying when she is ill. 5 My hints failed to impress Sally. 6. He can not tear himself away from TV. 7. He was drawn into a smuggling ring. 8. Camping trips require / call for hard work. 9. She is a woman of worldly knowledge and refinement. 10. These are really very elaborate and complicated instruments. 11. Your opinion is of great importance to me. 12. This apparatus is to be used only in case of emergencies that demand quick action. 13. "SOS" is a message requiring immediate action. 14. Do not pack the girl's head with fancies.
1. What will a mother feel if her child is late in returning? 2. What do you say when a room wants ventilating? 3. What kind of news will cause fear or anxiety? 4. What would you say of a woman of worldly knowledge and refinement? 5. What would you say of grandparents when they try to please their grandchildren in every way? 6. What would you call a person who Is in the habit of constantly watching TV? 7. What would you advise a person who is very particular about all kinds of little things? 8. What would you call a present that may not be expensive but is very dear to you?
the best in him. 11. They gave him an injection but it did not bring him .... 12. She wants to bring ... all the old customs.
1. Robinson Crusoe was seized with ... when he saw the footprint on the sand. 2. There is always a danger of ... when a theatre catches fire. 3. The thought that she might fail the examination filled her with .... 4. He lived in constant ... of his neighbours.
1. Some works of art have no ... for they are unique and, therefore, .... 2. The ... of a good education can not be measured in money. In Great Britain public education is free. It costs nothing. The ... of books may seem high, but their ... to a student who is educating himself may be great.
ми створюються для досвідченої (підготовленої) аудиторії. 19. Всі визнавали, що місіс Ерлінг мала витончений смак. 20. Ви надали мені неоціненну послугу. 21. Я ціную вашу думку більше ніж чиєсь інше. 22. Цінність деяких речей не може бути виміряна грошима. 23. Невідкладний виклик змусив лікаря відправитися до хворого в таку ніч. 24. Вона попросила не турбувати її, якщо не було нагальної потреби. 25 У мене голова забита всякими тривожними думками. 26. Індичка, зазвичай фарширована, - обов'язкова страва в американській сім'ї в День Подяки.
| 2019-04-18T15:21:13 |
http://um.co.ua/9/9-14/9-146270.html
|
0.999997 |
Generating natural behavior for synthetic humanoids such as virtual characters and humanoid robots in interactive settings is a difficult task. Many degrees of freedom have to be controlled and coordinated so as to realize smooth movements and convincing reactions to the environment and interaction partners. A common approach for the generation of human-like behaviors is based on motion capture data. Movements recorded from human demonstrators are replayed and possibly slightly altered to fit the current situation.
So far, however, approaches based on real-time adaptation of motion capture data have been mostly restricted to settings in which the behavior of only one agent is recorded during the data acquisition phase. As a result, the mutual dependencies inherent to the interaction between two agents cannot be represented and, thus, reproduced. During a live interaction, a synthetic humanoid may not be able to appropriately respond to the behavior of a human interaction partner.
Moreover, due to recent developments in cheap motion capture technologies, e.g. the Kinect camera, there is an increasing need for algorithms that can generate responses to perceived body movements of a human interaction partner. E.g. in gaming, a virtual character should recognize the behavior of the human player and respond with an appropriate reaction.
In this paper, we present a novel approach for realizing responsive synthetic humanoids, that can learn to react to the body movements of a human interaction partner. The approach extends traditional imitation learning [ BCDS08 ] to settings involving two persons. Using motion capture technology, the movements of a pair of persons are first recorded and then processed using machine learning algorithms.
The resulting interaction model encodes which postures of the passive interaction partner have been obtained depending on postures of the active interaction partner. Once a model is learned, it can be used by a synthetic humanoid (passive interaction partner) to engage in a similar interaction with a human counter part. The presented interaction learning approach addresses not only the question of what to imitate, but also when to imitate (c.f. [ DN02 ]).
This paper is organized as follows: In Section 2 we discuss related work. In Section 3 we introduce two-person interaction models and describe how to learn them from the simultaneous motion capture of two interaction partners. Section 4 shows examples how two-person interaction models can be applied to the behavior generation of virtual characters. In Section 5, we discuss two alternative machine learning algorithms for training our two-person interaction models. In section 6 we discuss the applicability of two-person interaction models to humanoid robots briefly before we finally conclude in Section 7.
Motion capture has been widely used for animating virtual characters. However, adapting the acquired motion data to new situations is a difficult task that is usually performed offline. In contrast, interactive settings as considered in our approach require the real-time adaptation of motion capture data. Some of the previous approaches for this are discussed in the following.
In [ YL10 ], for example, feedback controllers are constructed that can be used to adapt motion capture data sequences to physical pertubations and changes in the virtual environment. The approach formulates motion tracking as an optimal control problem whereby optimization methods derive parameters of a controller for real-time generation of full-body motions.
Multon and colleagues [ MKHK09 ] present a framework for animating virtual characters, where motion capture data is adapted in real-time based on a morphology-independent representation of motion and space-time constraints.
Ishigaki et al. [ IWZL09 ] introduced a control mechanism that utilizes example motions as well as live user movements to drive a kinematic controller. Here, a physics model generates a character's reactive motions.
Other work aims to adapt motion capture data through the inclusion of inverse kinematics solvers, e.g. [ KHL05, MM05 ].
Generating natural looking motions is also a focal point in the field of humanoid robotics. Researchers have extended the concept of motion capture by introducing imitation learning techniques, which can learn a compact representation of the observed behavior [ BCDS08 ]. Once a motion representation is learned, it can be used to synthesize new movements similar to the shown behavior while at the same time adapting them to the current environmental conditions. Imitation learning, thus, aims to combine the advantages of model-driven approaches, such as adaptability to unknown execution contexts, with benefits of data-driven approaches, such as synthesis of more natural-looking motions.
In an motion generation approach presented by Calinon et al. [ CDS10 ] Gaussian Mixture Regression models and Hidden Markov models (HMMs) are used to learn new gestures from multiple human demonstrations. In doing so time-independent models are built to reproduce the dynamics of observed movements. In [ YT08 ] the authors utilize recurrent neural networks to learn abstract task representations, e.g. pick and place. During offline training a robots' arms are guided to desired positions and the teachers' demonstrations are captured.
Figure 1. Use of two-person interaction models: Two humans' live-motions are captured, e.g. using the Kinect depth camera, and projected to a low-dimensional posture space. This space is used to learn a model of the shown interaction which is then utilized for motion generation for a synthetic humanoid in real-time.
In recent years, various attempts have been under- taken for using machine learning in human-robot-interaction scenarios. In [ WDA12 ], an extension of the Gaussian Process Dynamics Model was used to infer the intention of a human player during a table-tennis game. Through the analysis of the human player's movement, a robot player was able to determine the position to which the ball will be returned. This predictive ability allowed the robot to initiate its movements even before the human hit the ball. In [ IAM12 ], Gaussian mixture models were used to adapt the timing of a humanoid robot to that of a human partner in close-contact interaction scenarios. The parameters of the interaction model were updated using binary evaluation information obtained from the human. While the approach allowed for human-in-the-loop learning and adaptation, it did not include any imitation of observed interactions. In a similar vein, the work in [ LN10 ] showed how a robot can be actively involved in learning how to interact with a human partner. The robot performed a previously learned motion pattern and observed the partner's reaction to it. Learning was realized by recognizing the observed reaction and by encoding the action-reaction patterns in a HMM. The HMM was then used to synthesize similar interactions.
However, in all of the above approaches, the synthetic humanoid's motion is generated from motion demonstrations by only one actor. In contrast, the approach presented below draws on the simultaneous motion capture of two actors that demonstrate example human-human interactions.
In this section, a novel interaction learning method is introduced that allows virtual characters as well as humanoid robots to responsively react to the on-going movements of a human interaction partner (see Figure 1). At the core of our approach is an interaction model which is learned from example interactions of two human actors.
The interaction model is built in three steps: (1) Interactions between two persons are recorded via motion capture; (2) the dimensionality of the recorded motions is reduced; and, (3) a mapping between the two low-dimensional motions is learned.
In the first step, the movements of two people interacting with each other are captured. In general, the presented interaction learning approach is independent of particular motion capture systems. For development and testing, we used the Kinect sensor to record joint angles.
The precision of the calculated joint angles depends on the sampling rate. In general, low sampling rates lead to small datasets lacking accuracy whereas high sampling rates result in larger datasets with increased precision. Also, higher sampling rates may lead to redundant joint angle values for slow motions. For the behaviors used in this paper a sampling rate of 30fps is used.
Recorded motions are very high dimensional as 48 joint angles are captured per actor. A problem that arises when recording such motions is that not all measured variables are important in order to understand the underlying phenomena [ EL04 ].
Hence, dimensionality reduction should be applied to remove the redundant information, producing a more economic representation of the recorded motion. To this end, we use Principal Component Analysis (PCA) for construction of a low-dimensional posture space [ Amo10 ] which will be explained further in the following section.
Principal Component Analysis (PCA) is applied to the high-dimensional motion capture data to yield a low-dimensional posture space [ Amo10 ]. PCA reduces the dimensionality of a dataset based on the covariance matrix of modeled variables. Dimensionality reduction is achieved by finding a small set of orthogonal linear combinations (the principal components) of the original variables depending on the largest variance.
Previous results indicate that for most skills 90% of the original information can be retained with two to four principal components [ ABVJ09 ]. Each point in the posture space corresponds to a pose of the synthetic humanoid as illustrated in figure 2. Accordingly, a trajectory in posture space corresponds to a continuous movement. Hence, new behaviors can be created by generating trajectories in the posture space and projecting these back to the original high-dimensional space of joint angle values [ Amo10 ].
Figure 2. A low-dimensional posture space for a kick behavior reduced to two dimensions with principal component analysis. Every point in this space corresponds to a posture which can be reprojected to its original dimensionality and adopted by a virtual human or humanoid robot.
A two-person interaction model is the combination of two low-dimensional posture spaces with a mapping function from one posture space to the other. It is important that every point in the input posture space can be mapped to the output posture space and not just a subset of points, e.g.points lying on the low-dimensional trajectory of the demonstrated movement. In this way, the interaction model can also map postures that are similar to but not exactly like postures seen during training.
In the following, two machine learning algorithms with suitable generalization capabilities for learning such mapping functions are described.
Learning input-output relationships from recorded motion data can be considered as the problem of approximating an unknown function. A Feedforward Neural Net (FNN) is known to be well suited for this task [ MC01 ].
Figure 3. Multiple human postures are mapped on single robot/virtual human posture. Input to both nets (Left: FNN; Right: RNN) are the low-dimensional postures from a sliding window over time, i.e. the current posture with a history of preceeding postures. The output is the low-dimensional posture that represents the robot's/virtual human's response to the currently observed posture.
In our experiments, we use a simple FNN consisting of three layers, i.e. an input, a hidden and an output layer. How many neurons the hidden layer consists of and which connectivity value has to be used, depends on the recorded motion data. In general, the net should just have enough neurons to fit the data adequately while providing enough generalization capabilities for complex functions. We use 10 neurons in the hidden layer since we found that this fits the data adequately while retaining generalization capabilities for unseen low dimensional points. More neurons could increase generalization but also increase the risk of overfitting the data.
For the training of a FNN Levenberg-Marquardt backpropagation is utilized and all points from the low-dimensional trajectory of the active interaction partner are used as training data. Overfitting is avoided by using early stopping. For that low dimensional data is divided into two subsets: for training and testing.
The training data is used to adapt the nets weights and biases. The test data is utilized to monitor the validation error. This error normally decreases for a number of iterations. When it starts increasing again, this indicates overfitting and the training is stopped.
Generally, not only the current posture but also its history is necessary to determine an appropriate response. As FNNs have no short-term memory, the input to the FNN cannot be limited to the current posture. Instead, theFNN takes as input a sequence of low-dimensional postures, i.e. the current posture and a number of history postures. The labels (output) of the net are all low-dimensional points on the low-dimensional trajectory of the second, reactive interaction partner. Figure 3 shows how a sequence of postures is mapped onto a single output posture using an FNN.
Strict feedforward neural nets have no short term memory and cannot store history postures. Desired memory effects are only created due to the way how past inputs are represented in the net. Hence, a sliding window has to be used. And, thus, the amount of input neurons increases with each additional pose by the number of dimensions of the low-dimensional posture space.
To circumvent this problem we utilize layered recurrent neural nets (RNN) [ Elm90 ] which have proven to be well suited for modeling various temporal sequences [ FS02,YT08,NNT08 ]. RNNs have a feedback activation in the hidden layer where each hidden neuron feeds back to all hidden neurons. This embodies the desired short term memory. The hidden layer is updated not only with the current external input but also with activations from the previous time-step. This allows the usage of smaller sliding windows and, therefore, less input neurons. In our experiments the amount of neurons in the hidden layer is the same as for feedforward neural nets.
Figure 4. Two different punches and defenses have been used to learn an interaction model. Left: A low punch is defended with both arms pointing downwards. Right: A high punch is defended by pulling both arms up for defense.
For training, we use Bayesian regulation backpropagation. Training examples are low-dimensional points of the active partner's movements and their mapping onto low-dimensional points on the trajectory of the reactive interaction partner (see figure 3). The trained RNN defines a continuous mapping from the input posture space to the output posture space while preserving the temporal context of the recorded interaction.
In the following, we describe two examples where two persons are recorded while performing martial arts. The first interaction is a combination of punches varying in height with proper defenses. The second scenario consists of different kicks also with suitable defenses. For both behaviors, a virtual human learns to block the attacks with the trained defenses.
The motions of both the attacker and the defender start and end in an upright position with both arms stretched. The Kinect sensor was used as motion capture device in both examples.
In the first example a two person interaction model is used to generalize various punches and continuously compute motions for a defending virtual character. For that, two behaviors have been recorded. The first is a sequence of low punches at stomach level where the defender stretches his arms forward to block the attack. The second behavior consists of high punches at face level where the defender had to pull up both arms for a proper defense. In the center of figure 4 both punches with their defense motion are shown.
The training of the mapping algorithms is based on these two recordings. For each behavior, three repetitions of the respective punching style and their defense are captured. Then, the motions are projected into low dimensional space which can be seen in figure 4 for both interaction partners.
Figure 5. The image shows three punches for a user driven avatar (red skeleton) and the calculated virtual characters defense motion (blue skeleton). The user's motion on the left and right were similar to the ones used for model learning. However, the punch height in the middle has not been trained explicitly. Still, the learned model can calculate plausible motions.
All low dimensional points of the behavior trajectories are utilized to train both mapping algorithms.
In our experiments a 10 point pose history for FNNs, i.e. a 0.66 second window, has proven to be well suited. For RNNs good learning results have been achieved with only two points in the history.
For live-interaction with a virtual character, the user's current posture is captured at 30fps and projected into the previously created low-dimensional posture space. In combination with previous points a sequence of history poses is used as input for the mapping algorithm which predicts a low-dimensional point in the virtual character's posture space. The resulting point is then projected back to its original dimension and used as target pose for the virtual character.
Figure 6. Two different kicks and defenses were used to learn an interaction model. For low kicks the defender needs to crouch down in order to reach the attackers leg.
Figure 5 shows how the learned two person interaction model calculates the postures of a virtual character (blue skeleton) depending on postures obtained by a user driven avatar (red skeleton). The left and right of the figure display a newly recorded user motion similar to the ones used to train the mapping algorithm. However, the motion displayed in the center of the figure exhibits a punch height that has not been trained explicitly. Nevertheless, the interaction model can calculate suitable motions for the defending virtual character.
With the learned mappings, the virtual character can defend itself against the trained attacks. A clear differentiation between punching heights has been learned. Even for varying punch heights, the virtual human can still respond with correct defense motions.
Figure 7. A learned interaction model is used to calculate a virtual character's defend motions (blue character) based on a user controlled avatar (red character). The kick heights on the left and right are similar to the ones obtained during training. However, the kick height in the center was not present in any recording. Nevertheless, the two-person interaction model can generate suitable motions for the defending virtual character.
In the second example a two-person interaction model is learned from various kicks allowing a virtual character to defend itself with suitable defenses. For that two different kicks have been recorded. The first is a low kick where the defender needed to bend its knees in order to reach the attackers' foot. The second is a higher kick which was defended with an upright position (see figure 6).
For model learning both attack styles were performed three times and combined in a single animation. Since each motion started and ended in an upright position, an enclosed trajectory is created in low-dimensional space, as can be seen in figure 6.
This 4-dimensional posture space contains enough information to reconstruct animations with a negligible error. The neural nets were configured to consist of a hidden layer with 10 neurons. Addtionally, the size of the pose history has been set to 5 for FNNs and 2 for RNNs respectively.
For a live interaction with the character, the user's current posture is once again fed into the mapping algorithm and an appropriate pose for the defending character is predicted. Figure 7 shows a user controlled character repeating various kicks which where similar to the ones used to drive the nets training.
Additionally, a kick height (figure 7 center) not present in the recordings was performed. Since, varying kick heights of the attacker result in different low dimensional points, the virtual characters trajectories are located in the range of the previously recorded motion.
Figure 8. Euclidean distances are used to validate both mapping algorithm's qualities. Here, the graph shows distances for a novel punch data set. As it can be seen RNNs exhibit a smaller overall error.
For all three kick heights the virtual character learned when to defend itself using a crouched defense and when to block in an upright posture. In both cases the arms need to block the attacker's leg at the right time. For similar kicks, yet with different heights, proper defenses are generated by both FNN and RNN.
The interaction model that has been learned during this example allows the control of a virtual character based on demonstrations of only two variants of the behavior. The character is controlled continuously in a low-dimensional space and the model is robust to unseen user postures. Since the executed behaviors are based on human motion data, a life-like appeal of virtual characters can been produced.
The interaction models for the examples were learned with both mapping algorithms described above, i.e. FNN and RNN. In order to evaluate the quality of the mappings, the Euclidean distances between desired and trained low-dimensional points are measured, e.g. the training error. For that we recorded another punching example that is similar to the original human motion. A comparison of the resulting error is shown in figure 8 for the punching example.
Overall, we found that both mapping algorithms produce similar results. However, FNNs tend to exhibit weaker generalization capabilities. With varying interaction learning scenarios, differing net sizes were used in order to minimize the training error. Since the required size is hard to determine beforehand a cumbersome, trial and error process is inevitable.
Figure 9. A pose history improves the overall mappingquality. The figure illustrates how various sizes influence smoothness and response time for the punch behavior of a RNN. With increasing history sizes the net's prediction is shifted to the left, leading to larger response times but at the same time smoother motions are generated. We found that three points, i.e. the original point and two previous points in the posture history produced the best results.
Figure 10. Utilizing a two-person interaction model, a humanoid robot can react to the on-going behavior of a human interaction partner. The robot's motion is controlled by a low-dimensional mapping which preserves temporal coherences of the trained behavior.
In contrast, RNNs are more tolerant towards untrained inputs and no architecture changes have been required throughout the examples. Furthermore, RNNs maintain a short term memory, which renders posture histories unnecessary. However, we found that by the inclusion of a small number history poses, i.e. two additional postures, smoother responses are generated by the RNN. Figure 9 shows the prediction for training data with various history sizes. As can be seen, the overall response time of the synthetic humanoid increases if additional postures are used to predict the character's pose. We found that 2-point histories, i.e. the original point and two previous points, represent a good compromise between fast response times and high smoothness of the generated movements. This corresponds to the works of Ziemke et al. [ Zie96 ] which came to a similar conclusion. The combination of recurrence and a small sliding window tends to work best.
Two person interaction models can also be used to control humanoid robots, although an additional processing step to adapt the resulting motion is necessary. It is generally known that human motion cannot be transfered to robots without further optimization and is generally referred to as the correspondence problem [ ANDA03 ]. To overcome this the low-dimensional trajectory of the defender has to be altered to fit the robot's kinematic chain and stability constraints. For that we utilize an inverse kinematics solver.
The two-person interaction model thus maps the original low-dimensional movement of the attacker to the adopted low-dimensional movement of the defender (robot). The learned interaction model is then used to predict robot postures depending on observed user postures. As can be seen in figure 10 a Nao robot successfully mimics the demonstrated defense behavior.
In this paper we presented a new approach for teaching synthetic humanoids, such as virtual characters and humanoid robots, how to react to human behavior. A recorded example of the interaction between two persons is used to learn an interaction model specifying how to move in a particular situation. Two machine learning algorithms have been implemented to establish a mapping between (low-dimensional) movements of the two persons: Simple feedforward neural networks (FNN) and layered recurrent neural networks (RNN). While the FNN requires a large sliding window of current and recent body postures as input, the RNN can store previous poses in its internal state and thus can produce appropriate responses with a small sliding window of history poses. In our experiments, the RNN performed slightly better than the FNN.
In contrast to previous approaches to interaction modeling found in computer animation and game theory, no explicit segmentation of the seen behavior into separate parts is required. The responses of the synthetic humanoid are calculated continuously, based on learned mapping between the body postures of the observed humans. As a result, the interaction model can generalize to different situations while at the same time producing, smooth continuous movements.
Our method can be used to learn how to control synthetic humanoids from observation of similar situations between two humans. This can help to significantly increase the realism in the interaction between a robot and a human, or a virtual character and a human.
So far, however, our approach is based on joint angle data and, hence, response generation is based on similarities in joint-space. However, in order to more accurately reflect critical spatial relationships between the body parts of the interaction partners, motion synthesis could also be based on optimizations in task-space. By doing so, we could also better account for varying body heights of both interactants.
An interesting addition to the current approach would be a higher-level component for planning and strategic decision making. This could, for example, be realized through a combination of our approach and the approach of Wampler et al. [ WAH10 ]. We also hope to model multiple possible responses per stimulus. In our current setup each movement of the human can trigger only one possible response of the virtual character. A mixture-of-experts approach [ RP06 ], in which several interaction models are combined, can help to overcome this problem.
[Amo10] Heni Ben Amor Imitation Learning of Motor Skills for Synthetic Humanoids Technische Universität Bergakademie Freiberg 2010.
[ANDA03] Aris Alissandrakis Chrystopher L. Nehaniv Kerstin Dautenhahn Hatfield Herts Al Ab Solving the Correspondence Problem between Dissimilarly Embodied Robotic Arms Using the ALICE Imitation Mechanism Proceedings of the second international symposium on imitation in animals and artifacts, pp. 79—92 2003.
| 2019-04-25T16:26:13 |
https://www.jvrb.org/past-issues/11.2014/3856/view?set_language=de
|
0.999999 |
Is DRC's latest Ebola outbreak more challenging than the last?
Ebola has reared its ugly head again in the Democratic Republic of the Congo. The DRC has dealt with more outbreaks of the disease than any other nation. The battle to contain it comes amid violent conflict and political crises.
The Democratic Republic of the Congo recently announced the country's second Ebola outbreak in less than three months – the 10th since the virus was first identified near northern Congo's Ebola river. The news came barely a week after it declared the end to an epidemic in the northwest of the vast country which is also in the middle of political conflict.
The outbreak of the deadly virus has killed dozens of people in Ituri and the Beni region of the conflict-ridden North Kivu, bordering Uganda and Rwanda. North Kivu is the stronghold of a notorious Uganda-linked militia called the Allied Democratic Forces (ADF).
Results of genetic analysis from INRB #DRC just in: #Ebola Zaire confirmed, sequencing does not demonstrate close linkages with Equateur outbreak strain.
Ebola is back but what's different this time?
The Ebola "Zaire" strain is affecting North Kivu. Sequencing shows it is unrelated to the Equateur outbreak strain; meaning this is a new outbreak, WHO’s Deputy Director-General Dr Peter Salama said in a tweet.
Even “one laboratory-confirmed case of Ebola constitutes an outbreak,” WHO spokesperson Tarik Jasarevic told TRT World.
The 2014-2016 epidemic in was fought with help from the World Health Organization (WHO), which rushed in emergency aid, including protective gear, and unlocked $2 million (1.71 million euros) in fast-track financing.
So far 103 Ebola virus cases (confirmed and probable), including 63 deaths, have been reported, the WHO said on Twitter on August 24.
The outbreak is spreading through five health zones in North Kivu province and one health zone in Ituri province, WHO added.
How hard will it be to beat Ebola this time?
The situation could worsen as the affected areas host over one million displaced people, including refugees, and share borders with Rwanda and Uganda.
Refugees leaving the DRC's North Kivu province could spread the virus across borders.
Containing the virus might face bigger challenges due to the ongoing fighting between the government and militant groups.
Jasarevic told TRT World that the armed conflict makes it harder for WHO to respond to the outbreak.
“This makes the response very complex. North Kivu is UN security level 4, which is one of the highest levels of UN security phasing, and even though we have access to the towns of Mangini and Beni, we don’t know to what extent how we will best be able to contact tracing outside those small towns,” he said.
And cases have been reported from Beni as well, the WHO said on August 24.
At least 14 people kidnapped by the suspected ADF militia were found in Beni, army and civilian sources said on Tuesday, highlighting the region's volatility. The ADF, active in the east of DRC since 1995, is accused of massacring hundreds of people since 2014 by DRC and the United Nations mission in the DRC.
Response to Ebola in the region is further complicated by the presence of other armed groups and DRC army's unrestrained response, the New York University Study Group on Congo said in 2015.
“One of the many painful lessons from the devastating West African Ebola epidemic of 2014 was that the world expected much more from the World Health Organization than it was then able to deliver. Since then, we have ensured that the organisation is better prepared,” Jasarevic added.
Congolese officials and the World Health Organization began vaccinating health workers with an experimental vaccine on Wednesday, August 9, manufactured by Merck.
The vaccine which was first used in the tail end of the outbreak during the West Africa epidemic showed positive results when first trialled by the WHO and Doctors without Borders in the DRC in May. The 3,300 people who were vaccinated by the yet unregulated vaccine have as yet not developed Ebola.
“The current Ebola outbreak in DRC is caused by Ebola Zaire virus. The WHO Strategic Expert Advisory Group on Immunization (SAGE) recommended in 2017 that if an outbreak of Ebola virus disease is caused by the Zaire strain prior to the approval of the candidate vaccine, the rVSV-ZEBOV vaccine should be rapidly deployed under ‘expanded access’ or ‘compassionate use’, with informed consent and in compliance with good clinical practice,” Jasarevic said.
Ebola is a rare and often fatal illness caused by infections by one of five known virus species.
Ebola damages blood vessels which leads to a drop in blood pressure causing shock and multiple organ failure and can kill a person in one or two weeks.
Symptoms can include fever fatigue, muscle pain, headache and sore throat. Vomiting, diarrhoea, rash, symptoms of impaired kidney and liver function may occur. In extreme cases, the virus causes fatal bleeding from internal organs and orifices.
The symptoms start suddenly between two and 21 days after a person is infected. It is often difficult to diagnose Ebola as the disease is hard to distinguish from several other infectious diseases, including malaria, typhoid and meningitis.
Highly contagious, Ebola is contracted via direct contact with an infected person’s blood, secretions, organs or other bodily fluids as well as infected surfaces such as clothing.
Humans are infected when they come in direct contact with the blood, secretions, organs or other fluids of infected animals, including several primates, fruit bats, forest antelope and porcupines found ill or dead or in rainforests.
How did it all start in the Congo?
The latest outbreak of the haemorrhagic virus is the 10th in the DRC since 1976.
In 1976, the virus was discovered in the north of the country, then called Zaire, and named after a river nearby. That outbreak occurred in the Equateur province, with most cases occurring within 70 kilometres of Yambuku village.
It left 280 people dead out of the 318 infected, according to the Centers of Disease Control and Prevention (CDC), the US public health institute based in Atlanta.
The possible ground-zero was suspected to be Yambuku Mission Hospital where it was brought in from possibly Sudan where an outbreak had occurred. The disease was then transferred to the village population by unsterilised needles and other equipment.
WHO is working with local communities to contain the outbreak by providing vaccinations and the necessary information to residents on how to protect themselves from the virus.
Medecins Sans Frontieres is also working on the ground by setting up a 30-bed treatment unit in the town of Mangina as well as an isolation unit in Beni.
MSF is providing training on infection prevention and control to the health centres of the surrounding area.
The WHO said the organisation received over US$ 459 million in direct and in-kind donations from over 60 donors and most of this funding for Ebola-related activities came from governments and multilateral organisations.
The largest donors were the United States of America, the World Bank, the African Development Bank, the Ebola Multi Partner Trust Fund, and Japan.
| 2019-04-22T16:30:31 |
https://www.trtworld.com/africa/is-drc-s-latest-ebola-outbreak-more-challenging-than-the-last--19487
|
0.999999 |
Take this sentence: "The portable inkjet printer and wireless color photo printer included in XYZ enables ABC."
Should this be enable as you're referring to two different things, or is it enables because you are referring to two singular things?
Also, If I were to instead say" "The portable inkjet and wireless color photo printer", should printer be plural or singular there? I'm guessing if it stays singular then it would be enables and not enable.
Two things in a single unit are still correctly two things, and the two things "enable". But it is more common in American speech for people to follow the singular word, so far away from printers, with "enable".
I would say that "printers ... each enable" both is right and sounds right.
As for the alternative phrasing: from the user's point of view, not the mechanic's, is your unit essentially all one device, such as something that prints everything out the same port, or is it two devices in one stack or on one platform? That will tell you if it's singular or plural.
Should it be “100% donations policy” or “100% donation policy”?
Why do we say “ahead of” but not “behind of”?
| 2019-04-19T07:23:16 |
https://english.stackexchange.com/questions/489685/whether-to-use-enable-or-enables-after-mentioning-two-different-types-of-the
|
0.998911 |
While I was unplugging Blackberry from my PC running on Windows XP, I get an error message “A Media Card has been inserted that contains errors. To correct the errors please use a disk error-checking utility on a computer”. How can I avoid this error message?
It seems that your phone is corrupted and you have to run disk error-checking utility on your PC to resolve the errors. For that, you should first plug in the blackberry to your PC. Then, you should double-click on ‘My Computer’ option and select the appropriate disk (your phone) under ‘Removable Disk’ option. You should then right click on that disk and select ‘Properties’ option. A dialog box will appear on your screen after clicking on that option. You should select ‘Tool’ tab from it and click on the ‘Check Now’ button. After that, you should select the check box near ‘Automatically fix file system errors’ option and click on the ‘Start’ button. You should wait until the disk checking process finishes. Then, you should unplug your phone from PC and check whether the error message has disappeared or not.
your phone might get corrupted.
you've to run error checking utility to solve this problem.
plug the phone to your PC.
open my computer & select your phone.
right click on it to view properties option.
a dialog box will appear- in there you've to select the tool tab.
click on check now button.
select the checkbox associated with automatically fix file system error option & click the start button.
wait until this process finishes.
restart your computer to view the effect.
| 2019-04-23T18:46:12 |
http://windowsdiscussions.com/error-message-unplugging-blackberry-pc-running-windows-xp-1293.html
|
0.998274 |
Task: Write script for stories, as well as developing character personalities and profiles.
1) History of having written non-dialog stories before, such as fan fictions, scripts, etc. Genres of drama and adventure preferred over comedy.
2) Strong English language skills.
3) A very strong interest in this series and willingness to dedicate a lot of thought and passion into the characters and the story itself.
This is mostly up to you, but it shouldn't take more than from a few days up to a week, to finish.
Send your application to me by email (qvarsken (dot) mail (at) gmail (dot) com), discord (Advocat#6468), or simply post it below!
In your application, either in the body text or as a submission, be sure mention how you meet the required skills as well as either link or attach examples of your best work.
If you don't want to apply for writer but you still have many ideas for Minecraft animations; submit your ideas below!
Okay, thanks! I didn't really know where to post it.
| 2019-04-20T20:19:00 |
http://yttalk.com/threads/looking-for-a-writer-for-a-minecraft-animation.271560/
|
0.999972 |
WHY IS "EVERYONE" WAITING FOR SOMEONE ELSE TO DO IT?
I work with and listen to about two dozen radio stations. Thus, I am fortunate to be a part of quite a few very constructive planning, action timetable creation, brainstorming and goal-setting sessions each month.
Then, several months later, I return to those stations expecting to hear excitement, positive change and creativity on the air. Only, more often than I'd like, to be disappointed.
I recently helped a station create some exciting new jingles that cost us $5,000 to produce. Expensive, for the market size, but everyone who heard them thought they were worth it. Then, a month after the jingles were delivered, I visited the market and didn't hear a single jingle on the air in four hours of listening!
In another situation, following listener advisory panels that I moderated the station management team agreed that a 'no repeat workday guarantee' would be just what the listeners ordered.
Three weeks later, I checked in with the PD only to find out that were still "working out details" on implementation of the tactic. He wasn't sure when he could get it on the air.
I decided to write this item instead of either SCREAMING or FIRING SOMEONE, both of which I could easily do right now.
Speaking as a middle manager in one of America's larger radio companies, I wonder why it is that so few of our employees seem to see the sense of urgency in the opportunity in front of us. This is one of those high risk, high reward times that comes along only occasionally in every lifetime. Big changes are occurring; big demands are being placed on us. Big opportunity is ahead for those who are committed to rise to the occasion.
In radio today, there are - simply put - two situations: overwhelmed and unemployed. No question; many of us wish there was a THIRD option, but given the available choices, it seems that the wise selection is obvious.
I am not picking on anyone specific here, but I definitely see this as a general problem that many, many managers are talking to one another about. The bigger our corporations become, the more layers of bureaucracy there are and the more like characters in Dilbert many of us are becoming.
We can't allow that to happen to us. One broadcast paradigm has not shifted: radio is still the fast-moving, high-competitive, low loyalty business is has always been.
Thus, someone - our GMs, VPs, PDs, MDs, sellers and air personalities - has got to recognize the right thing to do and DO it, or we ALL pay the price.
Our owners are paying too much for our radio stations to allow us to become slower in our reaction times, complacent in our now-larger companies. We still only get seven or eight percent of the ad dollars and fewer than one fifth of the population is using radio in the average quarter hour.
Department of Justice concerns notwithstanding, our stations don't have a monopoly on anything and certainly NOT on our users' TIME. And, none of us can afford to act as if we do. The companies we work for are worth more than ever. That just means that we have a lot more to lose should we fail to please our listeners and media buyers.
THUS: when are we going to stop making excuses? There can always be a "good reason" why you couldn't do something. But, the people who manage to find a way in spite of all those good reasons are the winners. The ones with good reasons why they couldn't do something get left in high achievers' wake.
What can you do to minimize irritants NOW? Maybe the processing equipment isn't perfect. The spot load is too high. You're overworked. The listeners don't know those things. What can you do right now to make a difference they can HEAR?
"I don't think" needs to be replaced with "I am sure." If you don't know, find out. Don't guess. If you don't know what moves the meter among the many things you're being called upon to do, ask someone.
For example, what is REALLY important is there must be a clear definition of your position statement. Do you demonstrate what the words you say to describe your station to listeners MEAN to them, in terms they might use themselves? Do you have a user's guide to the radio station that is built to increase daily occasions of listening? Can you hear everything important that your station is about in a random half-hour of listening? Can this be executed flawlessly, every time, by even your weakest part-timer?
Ask part time, weekend and overnight talent if they understand precisely what the formatics need to be for your station to recruit listeners. If they don't know exactly what to do, that is not their fault. It is yours. Set up an easy-to-understand sound hour. When your liner says "new music," play new music. When a liner stages "variety," play variety. When a jingle says "fun," precede it with something creative and FUN.
Have the specific elements mapped out "bigger than life" in the control room so that it's easy and simple to do the format correctly. Make it difficult to do the wrong thing.
Your goal should be to ALWAYS have a better song on the air than the competition. On a routine basis, take several hours of each station's music that aired during the exact same time frame and refer to your research. How many times did our songs perform better than their songs?
If we are not hitting a 75% or better mark, then we should retool our library until it does! These are the basics of counter programming. When do they play their secondary tracks? Powers? Recurrents, etc? How about us? We should always be playing the better song, category for category, song by song.
Do the same exercise with current contest promos. Are you BIGGER, simpler to win, easier to play than THEY are? Is there an uncomplicated reason to listen built in? If not, what can you do to change that, immediately?
Keep "new stuff" coming on the air constantly. The stations that sizzle are constantly fresh and NEW. Never let the NEWNESS wear off your sound. This requires daily updating and rewriting of fresh, time-dated production imagery. It's not easy. But, it can be FUN. Specialize in SURPRISE. I cannot think of a better way to spend money than on the on-sir product and imaging is crucial.
Understand what elements are important and prioritize. Nothing moves the ratings faster than SIZZLE of HOT, FRESH and NEW unexpected, creative, well-produced FUN. Listeners know when you are faking it. And, speaking as "your" programming manager, I can too. So, please don't wait for me to discover it: DO IT. Develop non-negotiable standards for yourself and your coworkers. There is no person who is more impatient with the status quo than me. So, now that you know that (if you didn't already), use me as a resource. Together, we both can be agents of growth.
But, I can't do your job for you. And, making exciting radio is your job.
| 2019-04-25T08:46:52 |
https://danoday.com/articles/whyiseveryonewaitingforsomeoneelsetodoit.shtml
|
0.999859 |
1. Preheat barbecue to 350 degrees for indirect heating.
2. Place leak pieces in large bowl. Add onions, carrots, salt, pepper, mushrooms, sugar and oregano. Stir well.
3. In a cast-iron skillet over high heat, brown goose breasts in olive oil, until meat is seared on all sides. Set aside.
4. Pour vegetables over goose in skillet. Add half-and-half and stir well.
5. Cover top of skillet with pie crust, folding the edges under. Cut a few small vents in the crust.
6. Place over indirect heat on barbecue grill, cover, and cook for approximately 1 hour.
7. Remove from barbecue and serve directly from the skillet.
| 2019-04-19T16:34:47 |
http://www2.outdoorchannel.com/article.aspx?id=1746&articletype=article&key=Canada-Goose-Pie-Recipe
|
0.999999 |
Okay, now I'm really frightened-anyone have this?
I called my rheumatologist to make an appointment. I told the nurse about my sporatic inflammation of my knee, index finger and thumbs that come and go with no warning.
I wake up, and my finger is inflammed, very painful and will not bend or straighten. It is so inflammed the skin is shinny, however it is not warm or red.
The nurse made an appointment but told me to call immediately when it happens again so the doctor can see it. I asked her what it was and she said she didn't want to misdiagnose and say something incorrect but I needed to call the very next time.
Maybe she could hear the panic in my voice and she told me to calm down. I told her she really had me scared and worried by not telling me and I asked her if it was something scary. She tried to blow it off but I know there's something not right.
He ruled out RA, Lupus, MS and other forms of arthritis when he diagnosed the FM, but now I'm thinking I have something else too, since I get these "flares" every week or two.
Either my index finger or thumb will swell about twice it's normal size.
Can ANYONE help? I'm terrified and worried that it is something terrible.
My problems with my knee doing this have gone away since I've started taking Celebrex. It still flares, but the pain and swelling are greatly reduced.
For some reason my fingers and thumb are not effected by the Celebrex.
you answered my question.just thought Id mention that my mother in law use to be on celebrex for the arthritis in both her knees but it made them so swollen she couldnt walk so she had to stop taking it.plus she had also stated that you had to haver liver tests while you were on that.
Hi, I don't know what could be causing it but googled "finger swelling" and came up with 15 conditions it could be..?
fingers. I too get that. At times so bad that I can't open or close my hand. It feels like it's seized up, like it's rusty and needs oil. I've had Fibro so long, I don't ask too many questions anymore. I just say, "oh well another pain, another symptom and roll over to sleep some more.
My doctor thinks that I have a certain disorder, and from all the conversations that I've had with him, he honestly believes his diagnosis - but the labwork all comes back negative. He told me the next time I have an attack (which usually happen at night), I should go to the ER and tell them what tests to run. I say this in order to say that you shouldn't be so worried - sometimes doctors need to see exactly what you're describing to make a correct diagnosis. I hope this will help you calm down a little.
My rheumatologist told me to call the office right away if I get swollen, red, inflamed knees or ankles. With me, it's because all of the tests you can run have come up negative, I'm perfectly healthy on paper. However, I do get hot, red, swollen knees or ankles from time to time, which is why they originally thought I had RA. Unfortunately, the ONLY time this happens is at night, so none of my drs have been able to see me when my joints are swollen.
Ok, the good news is that they didn't say "Come in Immediately". THAT would freak me out. I agree with everyone else that sometimes the docs. need to see it happening to properly diagnose. Please keep us updated!!
I know from working at a doctors office a while back that they really can't tell you anything because they are not the doctor. try not to worry too much. I know it is easier said than done though LOL!!
Have you actaully been diagnosed with fibro? if you ahve then sorry for soundign liek a broken record, but those are all teh things they rule out during the three motnhs it takes to be diagnosed with fibro, there is no set test that will determine it just a set of guidelines.
I couldn't stay off of the internet yesterday doing more and more research.
I did get a fibromyalgia diagnosis and he said he ruled out everything else, however, I don't see on my paperwork where he tested for Rheumatoid Factor either time. Both my CRP and ESR rates were abnormal which showed inflammation.
My experience started when I was about 33 with random pain shooting down my arms, into my hands. I couldnt shake hands with people without wincing. My feet also were very tender. I was tap dancing in a stage show and found that my feet would start cramping and aching to the point that I would suffer great pain by the end of the show. I had no idea what was happening. I went to the doctor and it was suggested that I see a specialist in arthritis.
MY FIRST DIAGNOSIS WAS THAT I HAD FIBROMYALGIA which is another autoimmune disease that no one seems to understand why people have it. I DID NOT HAVE ANY PHYSICAL SIGNS OF RHEUMATOID ARTHRITIS at the time I went to the specialist.
No bumps, lumps, swollen wrists or ankles. THOSE CAME ONE YEAR LATER.
For a whole year I was taking Ibuprofen and trying to get sleep and seeing a therapist trying to relieve this strange disease Fibromyalgia, which as it turned out I didnt have. I have heard many stories of people who have been misdiagnosed in the beginning.
I FINALLY SAW A DOCTOR WHO DID A BLOOD TEST TO CHECK THE RHEUMATOID FACTOR, and also the erythrocyte sedimentation rate (ESR) which measures the speed at which red blood cells fall to the bottom of a test tube. The more rapidly your red blood cells drop, the more inflammation is present in the body. A HIGH SED RATE INDICATES THE INFLAMMATION AND the higher it is, the more severe the RA is.
X-rays may also be used over time to show the degeneration of joints and tissue though we did not take any at that time.
We found that the blood tests proved that I had rheumatoid arthritis.
| 2019-04-23T22:15:30 |
https://www.dailystrength.org/group/fibromyalgia/discussion/okay-now-i-m-really-frightened-anyone-have-this
|
0.999998 |
Why are glass doors so popular?
A well fitted glass door will ensure that the firebox is well sealed meaning heat cannot escape up the chimney or the flue as it would if the doors were left open. The 'chimney effect' is a phenomenon that occurs when warm air rises up through the chimney creating a suction and pulling more warm air from inside the house up with it.
This effect also causes cold air to be 'sucked' in through tiny gaps in windows and doors meaning that while the heat is escaping, the cold is coming in to replace it. This happens even while the fire is burning so investing in a blower , which can be used with the glass doors shut, may be worth considering. Using draught excluders may also be a good way to prevent a cold draught coming in.
Manufacturer instructions on fireplace glass door sets usually state clearly that the doors should be open when the fire is in use. In some cases it may be safe to use a gas fireplace with the doors shut to reduce the amount of cold air coming in through the chimney but this should always be checked with the manufacturer or a professional because there are situations where leaving the glass doors closed, while the fireplace is in use, could be dangerous. Most concerns are for consumer safety. If fireplace glass doors are used incorrectly or with the wrong type of fireplace, accidents could happen. Such situations include concerns over the use of incompatible products, for example, using air-tight glass doors with gas logs. Warnings on glass door products are there to protect both the manufacturer and the consumer so it is usually best to follow the instructions. Blowers can be installed to provide air circulation both inside the firebox and the room which enables glass doors be closed while using the fireplace.
Fireplace glass doors tend to absorb heat and so if the fire is burning with the doors closed, approximately two thirds of the radiation heat created will be absorbed by the glass instead of radiating out into the room. This means that fireplaces will still be more efficient with these doors open rather than closed, despite the potential heat loss caused by the afore-mentioned 'chimney effect'. When the fire begins to die down the doors can be closed to keep smoke out and heat in. Most glass door sets will come with a screen made from mesh to prevent embers or sparks from falling out onto carpet or flooring.
Glass doors for pre-fabricated fireplaces are usually fairly easy to measure and install as they tend to have standard fittings and measurements. Glass doors for masonry fireplaces, however, will be slightly different due to the irregularity of the stone fittings at the opening of the firebox. These will have to be carefully measured and glass doors are normally made to fit. It is best to get a professional contractor in for this. Fireplace glass doors also come in a variety of shapes with a wide range of style choices. The glass can be etched with designs, and even colored, to match the décor of the room.
| 2019-04-21T16:44:01 |
http://www.guide4home.com/dec-fire/glass-door.htm
|
0.999743 |
(SPOILERS) I wasn't among the multitude greeting the first Sicario with rapturous applause. It felt like a classic case of average material significantly lifted by the diligence of its director (and cinematographer and composer), but ultimately not all that. Any illusions that this gritty, violent, tale of cynicism and corruption – all generally signifiers of "realism" – in waging the War on Drugs had a degree of credibility well and truly went out the window when we learned that Benicio del Toro's character Alejandro Gillick wasn't just an unstoppable kickass ninja hitman; he was a grieving ex-lawyer turned unstoppable kickass ninja hitman. Sicario 2: Soldadograzes on further difficult-to-digest conceits, so in that respect is consistent, and – ironically – in some respects fares better than its predecessor through being more thoroughly genre-soaked and so avoiding the false doctrine of "revealing" how things really are.
There's been criticism that the picture’s mere existence feeds into the edifice (wall) of scaremongering regarding Mexico and illegal immigration. But equally, it would be an optimistic expectation that a movie focussing on the cartels would find the time to present a positive, balanced image of the country and its people. Added to which, Sicario 2 is expressly fuelled by the essential corruption of everything that is institutional USA, be it off-the-books activity sanctioned at the most senior level, incursions into foreign territory to enact illegal acts or hands-washing and loose-ends tying-off when operations don’t go quite as planned. Is it responsible to make something that may feed into the arguments of those pointing fingers? Probably not, and I'd be more willing to defend it against such a charge if the picture displayed an abiding intelligence or perspective amid the tried-and-tested pulp tropes.
The trigger for returning Taylor Sheridan’s screenplay is the Secretary of Defence's (Matthew Modine) desire for a persuasive response when cartel-smuggled suicide bombers lay waste a Kansas supermarket. The resultant suggestion from Matt Graver (Josh Brolin)? Start a war between cartels and so get them to wipe each other out. It’s a somewhat spirited-out-of-nowhere plotline, and one must assume it was a desperate attempt on the part of Sheridan to find a "What if?" that would lead to a politico giving carte blanche. Hence the later reveal that the suicide bombing is actually by American citizens. This might have been used as a blackly comic twist if it wasn't all-but lost in the melee, or even as a purposeful piece of misdirection à la WMD (something patently untrue used to justify an action), but since the operation is undercover that would only work as something specifically designed to manipulate Modine. There's a overriding air of jacked-up, Tom Clancy laziness about the device, but Sheridan isn't short on such shortcuts to plausibility.
The biggest of which being, in an age where even Hollywood is making movies about CIA drug running (Kill the Messenger, the recent Tom Cruiser Made in America), are we really supposed to buy that they'd go ahead with the instructions of the DoD and do something that might put a serious cramp in their ability to continue creaming off all those unofficial funding sources? Sheridan, in his bid for an authentic tone has omitted a key ingredient that would lend credibility to the cynicism of his anti-heroes.
As for those anti-heroes, he only makes matters worse for his title character by fleshing out Alejandro’s already ludicrous characterisation. A man who didn't hesitate to gun down a drug lord's wife and sons in the previous movie now gets all sentimental over another drug lord's daughter? Why? Because she evidently reminds him of his own dearly-departed bairn, and for the purposes of this instance can make the distinction between a drug lord and his offspring. As a consequence, he goes against his stone-cold code and Graver's instruction.
The biggest point in the character’s favour – and in Graver's – in the first movie was his completely amoral perspective (in terms of presenting him "legitimately"). Take that away, and you have just another surrogate father tale (Mercury Rising, Logan, Last Action Hero ad infinitum). Now to be fair, this plotline itself works reasonably well, but Sheridan is effectively telling us the guy we fully understood Emily Blunt wanting to kill in the original is alright, really. Does Sheridan actually think these guys are good guys? Despite everything they do? Is he someone destined to work closely with Peter Berg and Mark Wahlberg in the not-so-distant future? Well, he has family in law enforcement, so I shouldn't wonder.
This humanising extends to Graver by osmosis. He wouldn't usually hesitate to kill kidnapped drug lord's nipper Isabela (Isabela Moner, previously sexualised by Michael Bay in the triumphant Transformers: The Last Knight), but out of respect towards his dear friend, he decides not to. We’re now so far into the realm of Hollywood cliché that earlier turnabouts by Modine and admonishments from Catherine Keener seem entirely believable. On the other hand, who was director Stefano Sollima trying to kid from the first, when he has Alejandro reveal his face to a cartel lawyer on a busy street in Mexico City before killing him? How long can that kind of flagrantly frivolous vigilante/hitman make a fist of things? Outside of comic books, where he usually wears a mask.
There are other issues with the progression here; Isabela becomes largely passive (emotionally) once captured, a gauge of moral temperature rather than a character. Nevertheless, del Toro's such a good actor (and without a silly stammer here, even more so) that the bonding sequences are largely strong ones, particularly their encounter with a deaf peasant.
The picture's most laughable development comes after coyote – people smuggler – Miguel (Elijah Rodriguez) is required to execute Alejandro by boss Gallo (Manuel Garcia-Rulfo, probably resigned to playing despicable Mexicans, with this and Goliath in quick succession). Alejandro goes down, apparently lifeless. But we can’t help but notice he's been shot in the lower face, possibly only through his cheeks. Just what are the chances of that? Well, as I noted in my review of the original, Alejandro is pretty much Batman, so quite high in all honesty. It's precisely the sort of death-defying development that would befall a lawyer turned kickass ninja hitman. Come on, is anyone taking seriously the idea these movies are actually about something?
Sheridan even goes as far as setting up his would-be assassin as Son of Sicario or Sicario: The Next Generation (Blame it on Sicario?) for the inevitable sequel. In their favour, there's a sustained tension running through most of the Miguel scenes, a (relatively) innocent young guy dropped into people smuggling over his head, promised riches but likely to fall foul of border patrols and if not them a psychotic boss. Early scenes, such as his cousin trading insults with an acquaintance across the river/border as police car glides by, have an easy verisimilitude that's far more potent than the ominous perma-rumble of Hildur Guðnadóttir's swampy-industrial soundtrack.
Brolin's great in early scenes when unflinchingly called upon to kill a potential informant's family by drone strike, or meeting with Secretary of Defence and all but cracking open a beer as he enjoys the show and tells them how it is, but simply less interesting when it’s time to show he's also a compassionate, feeling kind of guy. Jeffery Donovan and Shea Whigham are wasted in a cameo and undercooked support respectively.
Sheridan justified his sequel with the thought "I sure would love to see what happened if these guys didn’t have a chaperone", going on to comment that Sicario was about the militarisation of the police and the sequel's next logical step was to remove that policing aspect. In both respects, he rather fails to find anything to dig into, neutering the potential of the leads unleashed by locating their moral centres and never building anything – in its predecessor either – into a coherent conversation regarding the latter. And yet, despite these fundamental problems, Sollima has made Sicario 2: Soldado a frequently gripping movie. Just don't call it authentic.
| 2019-04-22T12:41:49 |
http://www.nowinfullcolor.com/2018/07/dirty-is-exactly-why-youre-here.html
|
0.996614 |
BRNS is encouraging many institutions across the country to initiate research programmes on low-level radioactivity measurements. Other funding agencies such as the Atomic Energy Regulatory Board (AERB), Department of Science and Technology (DST), University Grants Commission (UGC), Ministry of Earth Sciences (MoES), etc., are also promoting research studies on radionuclides in the environment, and their applications as tracers, chronometer, etc. The new groups, which have recently initiated these studies, lack adequate infrastructure facilities (both measuring systems and calibration facilities) and expertise. In order to generate quality data, which can meet any stringent quality assurance programme, advanced equipment, calibration facility, and expertise are essential.
CARER is helping researchers from other institutions by extending its advanced facilities. During the last couple of years, users from other institutions have increased by manifold. Every effort is made to accommodate the request of researchers from different parts of the country.
CARER is accredited by the AERB for the measurement of radioactivity in commodities and issuing certificates as per the regulations of BIS, Govt. of India. Many industries from across India have approached CARER for testing of materials such as, spices, packaged water, pharmaceuticals, food materials, etc. meant for export as well as for public consumption. Besides this, many other industries such as, thermal power plants, mining industries, metal industries, etc. are using the facilities of CARER for testing their products for radioactivity. The availability of testing facilities at CARER has immensely benefited the industries of this region.
CARER is actively engaged in educating and creating awareness among the general public, college teachers, and students about radiation in the environment, the beneficial applications of radiation/radiation technology, and of the importance of nuclear power for the development of the country. In this direction, CARER is frequently organizing public awareness programmes in association with different colleges and institutions.
| 2019-04-26T12:21:48 |
https://mangaloreuniversity.ac.in/carer/support.html
|
0.997429 |
If we are an artists, and we've allowed others to read or hear or see our work, then we've probably had to deal with the pain of bad review.
We love our craft: our writing, our music, our poetry, our art, and our flow. We conceive it, we birth it, we bring it into the universe. And then we begin to share it with others. Along the way, we've probably gotten a lot of encouragement. Many of our friends have even enjoyed what they've seen and heard, and we are proud of our accomplishment. But then a "professional" steps in and calls our baby ugly.
It hurts, and sometimes it's downright depressing.
So how do we respond? What can we do? A lot.
1.We can reflect on why we were motivated to create. Did we do it for fame and glory or did we do it because we love the art? If we did it because we loved the art, then our love for it should diminish. If we did it for fame and glory, then maybe we need to make adjustments or get used to "mixed reviews."
2.We can examine the review closely and see if there's some truth to it. Do we have some growth as an artist? Is there something I could have done better? What can I learn from what the reviewer has said.
3.We can exercise some humility. Truth is, regardless of the quality of our output, not everyone will like our work. A review is an opinion, and we all know what they say about opinions: they are like noses (and various other body parts). On a side note, I was recently perusing Rotten Tomatoes and reading bios of some of my favorite actors. Most of my favorite actors have more rotten movies than fresh. As a matter of fact, according to RT, in order for a movie to be "fresh" or well-reviewed, only 60% of the reviewers have to like.
4.Don't forget that many artists aren't "critically" acclaimed but they are huge commercial successes.
5.Shake the haters off! Anybody can be a critic – it requires an opinion and the ability to string words together will some degree of proper syntax. And many reviewers enjoy giving negative critiques over positive one.
6.Finally, be wise! If several reviews point out the same weak point, listen! You're being told something. But if the reviews are a little of this and a little of that, maybe it's just a case of "different strokes for different folks." Hey! I don't think Kevin Hart is hysterical. But my girlfriend almost wets herself every time he walks on the stage.
Sharing your work is like sharing your heart. Getting a bad review is like sharing your heart and then having someone rip it apart! Just like love is a risk, sharing what's art is also. Whatever you do, though, whatever happens, whatever is said or reviewed, don't let it discourage you from pursuing those dreams.
Art is like fine wine; it gets better with age. If we refine our art, sift through the feedback and apply what's necessary, then our art will also get better with each stroke of the pen, the music, and the beat.
More in this category: « Are you inflating your social net worth?
| 2019-04-21T22:56:23 |
http://timothedavis.com/my-blog/item/102-the-pain-of-a-bad-review
|
0.998402 |
What does the legend of Vlad the Impaler have to do with the modern world? Is it possible that the Dracula of myth truly existedand that he has lived on, century after century, pursuing his own unknowable ends? Elizabeth Kostova's debut novel is an adventure of monumental proportions, a relentless tale that blends fact and fantasy, history and the present, with an assurance that is almost unbearably suspensefuland utterly unforgettable.
Late one night, exploring her father's library, a young woman finds an ancient book and a cache of yellowing letters. The letters are all addressed to "My dear and unfortunate successor," and they plunge her into a world she never dreamed ofa labyrinth where the secrets of her father's past and her mother's mysterious fate connect to an inconceivable evil hidden in the depths of history.
The letters provide links to one of the darkest powers that humanity has ever knownand to a centuries-long quest to find the source of that darkness and wipe it out. It is a quest for the truth about Vlad the Impaler, the medieval ruler whose barbarous reign formed the basis of the legend of Dracula. Generations of historians have risked their reputations, their sanity, and even their lives to learn the truth about Vlad the Impaler and Dracula. Now one young woman must decide whether to take up this quest herselfto follow her father in a hunt that nearly brought him to ruin years ago, when he was a vibrant young scholar and her mother was still alive.
What does the legend of Vlad the Impaler have to do with the modern world? Is it possible that the Dracula of myth truly existedand that he has lived on, century after century, pursuing his own unknowable ends? The answers to these questions cross time and borders, as first the father and then the daughter search for clues, from dusty Ivy League libraries to Istanbul, Budapest, and the depths of Eastern Europe. In city after city, in monasteries and archives, in letters and in secret conversations, the horrible truth emerges about Vlad the Impaler's dark reignand about a time-defying pact that may have kept his awful work alive down through the ages.
Parsing obscure signs and hidden texts, reading codes worked into the fabric of medieval monastic traditionsand evading the unknown adversaries who will go to any lengths to conceal and protect Vlad's ancient powersone woman comes ever closer to the secret of her own past and a confrontation with the very definition of evil. Elizabeth Kostova's debut novel is an adventure of monumental proportions, a relentless tale that blends fact and fantasy, history and the present, with an assurance that is almost unbearably suspensefuland utterly unforgettable.
In the "Note to the Reader," the narrator tells us, "There is a final resource to which I have resorted when necessary--the imagination." How does she use this resource in telling her story? Is it a resource to which the other historians in the book resort, as well?
The theme of mentors and disciples is an important one in the book. Who are the story's mentors, and in what sense is each a mentor? Who are the book's disciples?
Certainly there are many moments of tension and an overwhelming sense of impending dread, but Kostova doesn't stoop to the methodology used by so many modern thriller writers - short chapters, each ending with a breathless cliff-hanger forcing the pulse-rate up and, more often than not, the credibility down. Instead we get a challenging, substantial novel that many serious readers will love. (Reviewed by BookBrowse Review Team).
...Elizabeth Kostova has produced an honorable summer book, reasonably well written and enjoyable and, most important of all, very, very long: One can tote The Historian to the beach, to the mountains, to Europe or to grandmother's house and still be reading its 21st-century coda when Labor Day finally rolls around.
The writing is excellent, and the pace is brisk, although it sags a bit in the middle. There is plenty of suspense so that readers will want to find out what happens next. Ten years in the writing, this debut is recommended for readers who enjoy arcane literary puzzles la Dan Brown's The Da Vinci Code and Ian Caldwell's The Rule of Four.
Starred Review. Readers who think the legend of Dracula has become a trite staple of schlock fiction will find this atmospheric page-turner by first-time author Kostova a bloodthirsty delight. ...Both literary and scary, this one is guaranteed to keep one reading into the wee hours--preferably sitting in a brightly lit room and wearing a garlic necklace.
Starred Review. Exotic locales, tantalizing history, a family legacy and a love of the bloodthirsty: it's hard to imagine that readers won't be bitten, too.
Elizabeth Kostova (née Johnson) was born in New London, Connecticut in 1964, and now lives in Ann Arbor, Michigan. She graduated from Yale and holds an MFA from the University of Michigan where she won the Hopwood Award for the Novel-in-Progress. The Historian took her ten years to write, and was inspired by the vampire stories told to her by her father, a professor of urban planning, during the year they spent in Ljubljana, capital of Slovenia (where her father taught at the university) when she was seven, and as they traveled through Europe.
| 2019-04-23T02:52:59 |
https://www.bookbrowse.com/reviews/index.cfm/book_number/1589/the-historian
|
0.999091 |
The pod is always pretty much ready to go except for food and clothes and that's the way I weighed it. The dome, electric cords, chairs, hoses and some other weighty stuff is all in the aft locker. I may change that based on these results.
The propane tank is 3/4 empty.
I have not weighed the trailer yet, but do not believe we have more than 300# (hi estimate) in it which would put us right about 3,000#.
One surprising take away is how much that full tank of water affected tongue weight. In my case, it was an 80# difference or more than 20% difference in tongue weight.
So, that's my report. I'd really be interested in similar tests/results and any comments!
Not only interesting but useful. I don't, however, understand the significance of the bubble level being 1/8 down. If it had been level, would that suggest less weight on the tongue?
I need to do that with our 171 plus take it to a scale .. just to know.
The bubble level is my reference for tilting the trailer down in front a little (about an inch). This transferred about 10# more tongue weight. It should be exponential I believe. In other words, lowering the tongue 2 inches would transfer more than 20# weight to the hitch. Does that make sense?
The 268lb of stuff won't effect the tongue weight by 268lbs. The only way for it to do that would be if you stacked all 268lbs DIRECTLY on top the coupler.
Also if it's behind the axle as you said, it's going to effect it NEGATIVELY not positively. It has to be in front of the fulcrum (axle) to add to the tongue weight. But because the lever is shorter behind the fulcrum, weight place there has less effect than weight set equidistant from the opposite end of the lever.
This is why 240lbs of water only changed the tongue weight by 80lbs.
ARGGHH.. just reread that.. this 268lb is stuff in the TRUCK.. it also doesn't effect the tongue weight by 268lb from there, because it's weight is carried across the two axles of the truck. It does however count against your max GVWR. For tongue weight calculations, you strictly use the measured tongue weight of the trailer.
Based on that, I got the right size spring bars. If I went strictly by tongue weight, the bars I got would be way overkill for 300# tongue weight. Does that sound right?
With all the engineers here I'm surprised no one called me out on this!
Don't need no scale to measure tongue weight differences with water load. All you need is a freshman engineer nephew with an iPhone app or a tape measure!
Full tank (30 gal), TWF = 80.11584 lbs.
Funny how that is exactly what I measured. Ain't math grand!
Half tank (15 gal), TWF = 32.3048 lbs.
| 2019-04-25T10:42:14 |
http://rpod-owners.com/forum_posts.asp?TID=6228&PN=1&title=tongue-weight-redux
|
0.726909 |
Everyone has their own, special way of making potato salad so there are thousands upon thousands of variations. If you are used to pre-made, deli potato salad loaded with yellow mustard, this is far different.
Let me start by saying, I really dislike yellow mustard. I can't stand it. I will not eat a hamburger that has even accidentally had some wiped on it. So imagine how much I don't like to eat the average mustardy potato salad. Now on the other hand, I do like Dijon mustard and some of the more rustic, full grain mustards, and even dry mustard powder. I'm weird. I know.
This recipe has a bit of dry mustard powder in it so it's not completely bland. I would consider it to be a happy combination of mild deviled egginess and potato. I have taken this to various pot-lucks and had a number of people love it and ask for the recipe.
1. Scrub the potatoes. Combine the 1 Tablespoon of salt and potatoes in a large pot with enough water to cover them. Bring to a boil and simmer until tender. This should take about 40 minutes. If the skins start to split open, they are over-done.
2. After the potatoes have cooked about 30 minutes, start the sauce. In a 2 quart sauce pan, combine eggs, sugar, vinegar, butter, milk, dry mustard, and salt. Whisk to blend well. Bring up to a simmer over medium-low heat, stirring occasionally. Cook until thickened. Remove from heat and let cool while you continue.
3. Peel the potatoes while still warm. Be careful because they retain heat for a long time. Cut potatoes into chunks. Place in a large mixing bowl.
4. Take 3 hard boiled eggs and slice into nice rounds. Set aside. These will be used for topping the potato salad. Normally, I would use an egg/strawberry slicing device. However, mine broke the last time I used it to make egg salad. For very neat, thin, slices, I do recommend using some sort of slicer.
5. Dice the remaining eggs. Add to the potatoes in the mixing bowl.
6. Combine the slightly cooled sauce with mayonnaise. Mix well. Pour over the eggs and potatoes. Fold with a rubber spatula to evenly coat the potatoes without breaking up the pieces. Transfer to your serving bowl.
7. Arrange sliced eggs on top of the potato salad. Sprinkle with paprika. Cover and refrigerate overnight before serving.
Note: If you want crunchy onion and celery, add them to the sauce mixture in Step 6. If you want semi-crunchy onion and celery, add them to Step 2. Be sure to cook for 10 minutes.
If you want really soft cooked onion and celery, saute them in 2 Tablespoons of butter in a separate frying pan. Then once cooked to your liking, combine them with the other sauce ingredients in Step 2.
| 2019-04-18T10:51:46 |
http://www.tormentedkitchen.com/2011/07/jennys-eggy-potato-salad.html
|
0.999879 |
In "Through the Looking Glass", Alice uses a mirror to read which poem?
Who led the first expedition that reached the South Pole?
After watching US president Barack Obama giving a speech, David Letterman commented that it was nice to have a president who could pronounce which word?
Which of these is a female whale?
How many nouns are there in the following sentence: "Andrew gave his son a toy to play with"?
Where is "The Royal Tennis Court", a "real tennis" court built for Henry VIII of England who played there from 1528, that is still home to an active "real tennis" club?
Where are Southampton Island and the King George Islands?
| 2019-04-20T17:18:39 |
https://www.quizwise.com/general-knowledge-quiz/2017-02-15
|
0.99952 |
Can someone please help on how to handle tab separated records which have some tabs and newline characters in the data abinitio? I'm new to Ab Initio. One of my friends suggested to use RSV component, but I'm not able to figure out.
Examples of data would help, as would some file specification for it.
Tab separated records would be really unusual. It would lead to extremely long lines (as judged by the presence of newline (line feed) characters), which makes most utilities and editors choke eventually.
If tabs (or newlines) can be structural or data, then the data is indeterminate -- it is logically impossible to parse it. If you have some fixed-length fields, or some identifiable standard values, you might be lucky enoughg to pull out some logical structure with some degree of reliability.
Fundamentally, though, you have a bad file format, and you need to change whatever piece of junk is writing it this way.
Something like CSV uses commas and newlines for structure. It works round this issue using extra quoting mechanisms so you can identify what is "inside" a field and what is "outside" and therefore defines the structure unambiguously.
I remember posting to two threads on this site on similar questions, but I can't find the threads because Search is not smart enough. It only searches for known thread titles, not for contents of questions or replies.
"Having flat file where single row of data is in multiple lines. how to get a single row ignoring the white spaces"
and Search can't find that either. So I assume a lot of pre-conversion threads were just deliberately destroyed last year.
(1) OP had a file with field separators, but some process had folded the rows over multiple lines: that is, the data had extra newlines in most of the records.
Logical solution: we knew there should be 15 fields in each row. So we appended lines from the input file until there were 15 (or more) fields, and output that line. So the field separators told us where the real line breaks should be.
(2) OP had a CSV file where just one field could have extra commas in the data (but it had not been quoted as demanded by the CSV specification).
Logical solution: There were always 13 data fields, and field 4 was the only one that had extra commas (it was an address). So we identified fields 1-3 by counting from the front, and fields 5-13 by counting from the end of each line, and field 4 was anything in the middle, and we quoted it to make valid CSV. In that case, the line breaks told us where the trailing fields should be.
If your file format has that much internal consistency, we can fix it. If it really has random characters, nobody can fix it, because information has been destroyed when it was created.
Thanks. That was a great help for me.
In our case, the data is randomly distributed with the newline characters.
We have asked upstream team to replace the tabs and newline characters with spaces.
| 2019-04-20T21:25:37 |
https://it.toolbox.com/question/how-to-handle-tab-separated-records-which-have-some-tabs-and-newline-characters-in-the-data-ab-initio-031219
|
0.999075 |
A long, long time ago, a movie was made that celebrated the bravery of Naval aviators. It was produced by a famous film duo of the time - Jerry Bruckheimer and Don Simpson - who were known more for making big-budget action-packed films than for creating sublime pieces of cinematic art. The movie Top Gun, which they produced in 1986 with Tony Scott directing, definitely fits the formula one expects of a Simpson-Bruckheimer film. It stars Tom Cruise as "Maverick," a young upstart with an attitude problem and a lot of raw talent, whose unwillingness to play by the rules and work as a member of a team cause a lot of tension among his group of fliers. Mistakes happen, tragedy befalls a friend, and Maverick comes out a better man for it. But he still has that attitude problem.
The film did extremely well at the box office, becoming the top-grossing film of 1986, and an excellent recruiting tool for the U.S. Navy. In fact, when Paramount Pictures offered the Navy the chance to put a recruiting commercial on videocassettes of the film upon its release, the Navy declined, stating they didn't need that since the entire film served as a recruiting device. Critical praise was a little shortcoming for the film, but that did not really matter considering the money it made. With the untimely death of Don Simpson in 1996, and Tony Scott's suicide in 2012, discussions for a sequel to the film have slowed down, but persist. The American Treasure Tour celebrates this film with a plaque, located in our Toy Box.
QUESTION: Which Academy Award did Top Gun receive?
TUT TUT. Our continuing celebration of America and American history brings us to this day in Egypt during the year 1922, when the British archaeologist Howard Carter discovered the tomb of Pharoah Tutenkhamen in the Valley of Kings, a spot in the deepest desert where ancient leaders of Egypt buried their greatest citizens. This discovery had a far-reaching impact. As a previously undiscovered site, King Tut's gold had never been ransacked before. People around the world could see how a Pharoah lived and died, and mysteries of this ancient culture were answered. Then, of course, Steve Martin made a wonderful song about it half a century later.
DORIS! Turning ninety today, Doris Roberts has had a long and distinguished career in television, film and on stage, often playing a manipulative mother figure. She won a number of awards for her portrayal of Marie Barone, the dysfunctional mother in Everybody Loves Raymond. We're glad to see she's still going strong. Keep it up, Doris!
ANSWER: D) Best Music, for the song "Take My Breath Away."
| 2019-04-26T02:12:57 |
http://americantreasuretour.com/blog/top-gun-wednesday-november-4-2015
|
0.999846 |
An appellate court in Contra Costa County, Calif., has discovered a previously unknown pastime: cellphone joyriding.
It looks a lot like theft.
It involves a stranger taking violent possession of an owner's phone against the owner's will. It involves the owner being punched repeatedly in the head by the stranger. But, found the court, this isn't theft. It's "temporary taking"--something far more akin to a kid borrowing your car for a joyride.
"On December 20, 2010, Matthew Cardoza was standing outside the hospital where he worked, taking a 10 minute break. He was using his new Apple iPhone 4 cell phone to exchange text messages with his fiancee."
He had just hit "send" on his last text when up came an agitated man--a stranger to Cardoza--who said, according to the court filing, "Hey, man, let me get that phone."
The stranger took the phone out of Cardoza's hands and made off with it.
Cardoza gave chase, grabbed the man by the back of his shirt, spun him around and grabbed for the phone. They struggled.
During the struggle, the stranger yelled, "Give me the phone, give me the phone. I'll hurt you. Give me the phone." He punched Cardoza in the head several times.
Cardoza got the phone back, after which the man said, "I'll pay you, I'll pay you. I just want the phone." Cardoza told the man he didn't want his money and asked him please to leave him alone. The man tried to apologize and told Cardoza he needed to get help because his mother "had just got her house on fire."
The violent stranger was Kurt Carr.
It had already been a full day for Carr before the incident with Cardoza. That same morning, the mother of his fiancee had been driving him to a doctor's appointment at the hospital when the two had gotten into an argument. Carr decided to jump out of the car while it was still moving, according to the filing.
"He felt he had to call his fiancee immediately to tell her about his fight with her mother," according to court papers. He already was agitated and distraught because, according to the court filing, his mother's house had burned down a few days previously, "and he believed the fire was caused by faulty repair work he had done."
Carr went looking for a phone, and saw Cardoza on his new iPhone. "Man, let me get the phone. Let me use your phone....Can I use your phone?" he said to Cardoza, according to the filing.
Cardoza told him no. Karr asked again, and was again refused. The struggle for the phone then ensued.
The court found that in order to prove theft, the prosecution would had have to have shown that "[Karr] intended to deprive [Cardoza] of [the phone] permanently or to remove it from [Cardoza's] possession for so extended a period of time that the owner would be deprived of a major portion of the value or enjoyment of the property."
As for "permanently," the California Supreme Court previously had decreed, according to the appellate court, that permanently didn't have to be all that permanent. "Our Supreme Court has ruled...that this intent requirement is flexible and not to be taken literally."
As for "for so extended a period of time that the owner would be deprived of a major portion of the value or enjoyment of the property," said the court, the major-ness of the deprivation depends on circumstance.
"Taking cut flowers from a florist without consent, with intent to return them in a week" and "Taking a neighbor's lawn mower without consent for the summer, with intent to return it in the fall" would both qualify as theft, because the deprivation was major, the court wrote in its decision.
"Much depends," said the court, "upon the nature of the property and its expected useful life...It is one thing to take another's fresh strawberries with the intent to return them two weeks later, another thing to take his diamond ring with a like intention."
Which brings us to joyriding.
Joyriding--the temporary taking of another's car--is not theft, according to the court, because the deprivation is only temporary (and the car is non-perishable). Likewise, one kid's hiding another kid's bicycle overnight, in order to "get even" for the owner having teased the hider, is not theft but instead temporary taking.
Returning now to Cardoza and his iPhone, the court found that, "the phone was neither perishable in nature nor good for only seasonal use." Further, that Karr's taking of the phone was temporary, done in order for him to make a call and not to deprive Cardoza permanently of its use (nor to deprive Cardoza of a major portion of its value and enjoyment).
"Taking a phone for temporary use is far more akin to 'joyriding' or taking a bicycle with intent to return it the next day, which do not constitute theft," found the court.
A Superior Court jury acquitted Carr of robbery, but convicted him of grand theft. He received a sentence of eight months of home confinement. The appeals court unanimously reversed that decision and vacated the sentence.
| 2019-04-20T10:39:15 |
https://www.yahoo.com/gma/calif-appeals-court-tosses-iphone-theft-charge-temporary-031225849--abc-news-money.html
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.