Connect with us

Technology

Atomos Space’s first orbital mission is a trial by fire

Published

on

Few missions more vividly embody the maxim “space is hard” than Atomos Space’s first demonstration mission, which the corporate managed to drag back from the brink of disaster – greater than once.

This demonstration mission, called Mission-1, was launched into orbit on a SpaceX Falcon 9 rocket on March 4. The mission’s goals are extremely ambitious: the 2 spacecraft – an orbital transfer vehicle called Quark-LITE and a goal vehicle called Gluon – will ultimately reveal extremely complex maneuvers, including rendezvous, docking, orbital transfer and in-orbit refueling.

The company faced two major problems related to communication and the rotation rate of the spacecraft – and (largely) solved each problems despite massive limitations, sparse data packets, and very limited bandwidth. (So ​​limited, in reality, that the team needed to limit flight software updates to a text string of just 145 characters).

“It was relentless,” Atomos CEO and co-founder Vanessa Clark told TechCrunch.

William Kowalski, COO and co-founder of the corporate, agreed. “What makes it so difficult is that even in our situation, we’re trying to extrapolate the status of a very complicated system from maybe 100 bytes of data,” he said. “That’s a lot. You guess what’s causing it, knowing that a few of those guesses could lead on you down a path from which you may never get well.”

Problems began just hours after the 2 interconnected spacecraft were launched from the Falcon 9 upper stage. The deployment was nominal, and Atomos received the first signal from the spacecraft seven minutes after deployment. The mood was solemn.

But 40 minutes passed before the corporate received one other signal. Then eight hours.

Atomos expected data packets every jiffy.

“The worst (day) was Monday, when we took off, that evening,” Kowalski said. “It was 11 p.m. at night, it was me and the chief engineer… and we didn’t hear anything and we just think: Have we failed? Did they die? We gave it a shot, but it just didn’t work. It was really a punch in the gut.”

Mission controllers didn’t discover the basis cause until 24 to 48 hours after deployment, and did so with the assistance of one other company with on-orbit assets. After pulling some strings, they managed to speak on the phone to the chief systems engineer of the satellite communications company Iridium. The spacecraft used third-party modems that used the Iridium intersatellite link network and likewise used the Iridium constellation as relay satellites. The Atomos spacecraft was moving too fast and in direct contrast, it couldn’t perform a data “handshake” with the Iridium satellites to truly transmit the knowledge back to Earth.

Atomos engineers implemented a series of software updates that reduced duty cycles and ensured that the radios would all the time be on, even when the spacecraft was in a low-power state.

When engineers tried to resolve the communication problem, nevertheless, they encountered one other problem: the spacecraft was rolling at an especially high rate of 55 degrees per second (they were designed to deal with roll rates of as much as 5 degrees per second). In addition, the spacecraft slowly rotated in order that the solar panels not faced the sun. This meant it was a race against time and the spacecraft’s batteries completely depleted.

“We had two charts,” Kowalski said. “We plotted our power trend for when we predict we will probably be facing away from the sun and have (at) zero power, in addition to the sink rate. The removal rate needed to be delivered to zero before the ability dropped to zero.

The problem was exacerbated by limited communication; teams weren’t in a position to definitively confirm that anything was mistaken until the fourth day after deployment, and the spacecraft could only process recent commands between long periods of what were essentially communications blackouts.

Slowly, over the course of several days, they managed to slow the spacecraft down. The team achieved one other major victory after they were able to ascertain high-bandwidth communications, a space-to-space link on a Quark-LITE device communicating via the Inmarsat network. On Thursday, the corporate made its first attempt at establishing broadband connectivity and successfully maintained communication with the spacecraft for six minutes.

During this era, mission controllers received 17 times more data than since launch. As a result, mission controllers received enormous amounts of information on the state of the spacecraft. The news wasn’t all positive – certainly one of the OTV batteries was badly damaged by aggressive cycling and it appears the GPS aboard certainly one of the spacecraft needed to be reset – but these are easy fixes, Clark said.

The company plans to begin commissioning the drive system on Tuesday or Wednesday. If all goes in response to plan and engineers determine that the support system provides accuracy and aiming control, they are going to test operation without torque bars and response wheels. The company intends to deploy the spacecraft inside about a month, with all mission objectives expected to be achieved by the top of June.

Kowalski and Clark attribute a part of the startup’s success to the incontrovertible fact that it is highly vertically integrated. The team, which worked 100 hours per week within the first week after deployment, was in a position to use their in-depth knowledge of spacecraft design to resolve emerging problems.

“It was obviously very painful, but it is reminiscent of the words of the CEO of Nvidia: ‘I wish you great suffering.’ We went through it and it wasn’t great at the moment, but now that we’ve gotten there, we’re definitely better,” Clark said.

This article was originally published on : techcrunch.com
Continue Reading
Advertisement
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Technology

Data center technology is exploding, but implementation won’t be easy for startups

Published

on

By

Data center, data centers, data center tech

The data center industry is rapidly evolving to maintain pace with the event of artificial intelligence. While these data centers are essential AI infrastructure, they store the AI ​​company’s computations, are expensive to construct, seemingly costlier to take care of, and likewise devour enormous amounts of energy. Startups intend to make data centers more efficient and sustainable, but it isn’t that straightforward.

The value of the worldwide data center market is estimated at USD 301 billion, based on P&S Intelligenceand this market is expected to greater than double by 2030, reaching a market value of $622.4 billion. According to estimates, data centers currently devour roughly 4% of total power within the US Electricity Research Instituteand is projected to greater than double by 2030, to 9%.

Data centers and the massive firms that depend on them are fighting for power. Last month, Microsoft signed a take care of Constellation Energy to restart its nuclear reactor Three Mile Island to maintain up with demand.

As the number of information centers grows, so does the variety of startups trying to solve the energy and environmental crisis in the info center industry. Startups like Incooling and Submer need to address this space by cooling existing data center technology to supply less heat. Others, like Phaidra, use software that helps data centers manage cooling more efficiently.

Some wish to construct a very latest model. Verrus is constructing a more “flexible” data center using microgrids. Sage Geosystems is working on a method to use pressurized hot water to power data centers as an alternative of natural gas.

Sophie Bakalar, a partner at Collab Fund, an investor in Phaidra, told TechCrunch that while before the AI ​​boom there have been entrepreneurs trying to construct technology for data centers – data centers also play a big role in cloud computing and bitcoin mining – she noticed a 10-fold increase a rise within the variety of founders trying to construct technology for this space over the past 12 months.

“We saw a company that builds data centers in space, they manage the whole thing,” Bakalar said. “Whenever there is such an obvious supply and demand problem, it is natural to see many entrepreneurs willing to tackle the problem from different angles.”

While data centers are growing rapidly and can need solutions that provide greater efficiency, that does not imply startups should think that adopting their technology will be easy.

Data center challenges

Francis O’Sullivan, managing director at S2G Ventures, told TechCrunch that the speed at which this space is growing can actually make it harder for startups to search out partners willing to check or take a likelihood on their technology.

“(Data centers) are extremely expensive assets, multi-billion dollar facilities. They actually have to prove themselves,” O’Sullivan said. “That’s why the real, meaty world of data centers is not a forum for experimentation.”

The customer base for any such technology is likely more concentrated and subsequently likely harder to penetrate, said Kristian Branaes, partner at climate-focused VC firm Transition. Branaes added that his company spent plenty of time researching and diving into the info center technology category, but while it found some cool firms creating cutting-edge technologies, it wasn’t in a position to gain enough conviction to speculate.

Branaes worries about how firms will be in a position to scale. He thinks a few of the startups he’s found fit into the classic query about climate tech, which is that they are cool technologies, but they don’t seem to be necessarily firms that may generate risk-level returns. He said it’s difficult to construct a venture-scale company that sells its products to only a handful of huge firms like Microsoft and Apple.

“We have come to (this) view: It is very, very difficult to build a large company that only sells AWS and Microsoft and everyone else; they are ruthless in their purchases,” Branaes said. “They should not within the business of giving big margins. If you begin making an excessive amount of money, they’ll wish to work around it or start doing it internally.

Turning on

While some investors remain skeptical, many startups operating on this space are proving popular. Impending regulations in each cases Europe and in data center-heavy US states equivalent to Virginia mean that even when these large customers should not purchasing solutions now, they’ll likely have to accomplish that in the longer term.

Helena Samodurova, co-founder of Incooling, a Dutch startup trying to cool data centers, founded her company six years ago, before the present artificial intelligence hype. While data centers and the energy they devour were a priority back then, the necessity for Incooling technology has completely modified.

“People didn’t really know about it back then,” Samodurova said. “The situation has changed enormously over the last six years. As we went through this journey, we really had to educate people on what it was. Fast forward six years later, but that’s not the case. We are wanted.”

Samoduorva said there was increased interest from each potential customers and investors. She added that the info center industry is broader than simply the Amazons and Googles of the world and that helping reduce data center emissions is not only focused on these few large firms.

“You have a bus that you can take to the station, you have a car that you can take your family on a trip, you have a Ferrari that you can race on, everything has four wheels, but the mechanics are different,” Samodurova said. “We provide cooling solutions or computing solutions to solve any bottlenecks.”

O’Sullivan said that while for him, much of the info center technology is too young to get enthusiastic about, there are other categories of firms value supporting which might be helping to resolve a few of the same problems that data center technology is in search of to handle. One is to resolve the issues of getting the actual power into the info center and ensuring the ability grids can handle that level of power.

For data center-focused startups, adoption may simply be too early for a few of the category’s early entrants. Unlike Incooling, many firms have only been established in the previous couple of years. While the info center technology market may be in its infancy, the bogus intelligence and data centers needed to power the industry should not going away anytime soon.

“I think the main thing to consider is that this is really urgent,” Bakalar said. “The development is really outpacing the current infrastructure that we have. We need newer, better and faster ways to achieve the promise we have heard about artificial intelligence.”

This article was originally published on : techcrunch.com
Continue Reading

Technology

In the latest move against WP Engine, WordPress is taking control of the ACF plugin

Published

on

By

stylized wordpress logo

The dispute between WordPress founder Matt Mullenweg and hosting provider WP Engine continues, with Mullenweg announcing that WordPress “forks” a plugin developed by WP Engine.

Specifically, Advanced Custom Fields — a plugin that makes it easier for WordPress users to customize edit screens — is being taken out of WP Engine’s hands and updated as a brand new plugin called Secure Custom Fields.

Mullenweg wrote that this step was obligatory “to remove commercial add-ons and fix a security issue.”

Advanced custom fields team replied on X daydescribing it as a situation where a plugin “in active development” was “unilaterally and forcibly taken away from its creator without consent”, which he claims has never happened “in WordPress’ 21-year history.”

“This important community promise has been violated, and we ask everyone to consider the ethics of this action and the setting of a new precedent,” the ACF team wrote.

Both the Mullenweg blog post and answer from WordPress claims that similar situations have occurred before, although Mullenweg added: “This is a rare and unusual situation caused by WP Engine legal attacks. We don’t anticipate this will happen with other plugins.”

They also pointed to the WordPress plugin guidelineswhich provides WordPress the right to disable or remove any plugin, remove developer access, or change a plugin “without developer consent, in the name of public safety.”

A bit of background: WordPress is a free, open-source content management system utilized by many web sites (including TechCrunch), while corporations like Mullenweg’s WP Engine and Automattic moreover offer hosting and other industrial services.

Last month, Mullenweg published a blog entry criticizing WP Engine as “WordPress cancer”. His criticisms included the whole lot from WP Engine’s lack of version history support to its Silver Lake investor, but he also suggested that its “WP” branding confuses customers by making it appear as if the company is officially connected to WordPress.

Cease and desist letters went each ways, with WP Engine claiming that Mullenweg threatened to take a “nuclear scorched earth approach” if the company didn’t pay for the WordPress trademark license.

WordPress blocked WP Engine from accessing WordPress.org, briefly lifted the ban, after which re-imposed it. This essentially prevents WP Engine from updating the plugin via WordPress.org – so it might probably’t offer automatic updates to handle security issues.

WP Engine does, nonetheless posted a workaround for users who need to update the plugin and proceed using ACF. (It says this workaround is only obligatory without spending a dime ACF users, as skilled users will proceed to receive updates through the ACF website.)

Going forward, Mullenweg wrote that Secure Custom Fields can be available as a non-commercial plugin: “If any developers want to get involved in maintaining and improving it, please contact us.”

This article was originally published on : techcrunch.com
Continue Reading

Technology

Meta’s Yann LeCun says concerns about an existential threat from artificial intelligence are ‘complete nonsense’

Published

on

By

Artificial intelligence pioneer Yann LeCun doesn’t imagine that artificial intelligence is on the verge of becoming intelligent.

LeCun – Professor at New York University, senior researcher at Meta, and winner of a prestigious competition AM Turning Award — for instance, he has spoken openly about his skepticism before tweeting that before we worry about controlling superintelligent AI, “we need to have a beginning design for a system that’s smarter than a house cat.”

He developed his opinion in interview for the Wall Street Journal.where he answered a matter about whether artificial intelligence has turn out to be intelligent enough to pose a threat to humanity by saying, “You’ll have to excuse my French, but it’s complete bullshit.”

LeCun argued that today’s large language models lack some key cat-level abilities, equivalent to persistent memory, reasoning, planning, and understanding the physical world. In his opinion, LLMs only show that “you can manipulate language and not be smart” and won’t ever result in true artificial general intelligence (AGI).

It’s not that he’s a whole AGI skeptic. But he said latest approaches can be needed. For example, he identified that his AI fundamental research team at Meta is working on analyzing real-world video.

This article was originally published on : techcrunch.com
Continue Reading
Advertisement

OUR NEWSLETTER

Subscribe Us To Receive Our Latest News Directly In Your Inbox!

We don’t spam! Read our privacy policy for more info.

Trending