Connect with us


Tibbits are colorful pre-programmed modules for building IoT devices – TechCrunch



4 mins ago


At first glance, Tibbits look like building blocks, but each one is a module or a connector that makes it easier to build connected devices and systems. Tibbits were created by Tibbo Technology, a Taipei-based startup that exhibited at Computex this week (it showed off a humanoid robot built from various Tibbits).

Pre-programmed Tibbit modules from Tibbo

The heart of the Red Dot Award-winning Tibbo Project System (the company used bright colors to make its modules stand out from other hardware) is the Tibbo Project PCB, which includes a CPU, memory and Ethernet port. Then you pick Tibbits, with pre-programmed functionality (such as RS232/422/485 modules, DAC and ADC devices, power regulators, temperature, humidity or pressure sensors or PWM generators), to plug into your PCB. Once done, you can place your project in one of Tibbo’s three enclosure kits (custom enclosures are also available).

Tibbo also offers an online configurator that lets you preview your device to see if it will work the way you want before you begin building, and its own programming languages (Tibbo BASIC and Tibbo C) and app development platform.

Source link

Check Also

If you’re feeling flush this week, then perhaps instead of buying a second Bugatti you …

Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *


Apple pushes out iOS 14.2.1, and it’s mostly bug fixes



Enlarge / The iPhone 12 mini. iOS 14.2.1 fixes an issue that affected the lock screen on this phone.

Samuel Axon

Yesterday, Apple released iOS 14.2.1 to fix bugs users have encountered on new iPhones since iOS 14.2 launched on November 5. Unlike many other iOS releases, this release was not accompanied by updates to all of the company’s other operating systems.

The update fixes a bug that caused an unresponsive lock screen specifically on the iPhone 12 mini, and it addressed an issue that prevented MMS messages from coming in. Further, 14.2.1 fixes a problem with sound quality on connected hearing devices.

Here are Apple’s release notes for iOS 14.2.1:

iOS 14.2.1 addresses the following issues for your iPhone:

  • Some MMS messages may not be received
  • Made for iPhone hearing devices could have sound quality issues when listening to audio from iPhone
  • Lock Screen could become unresponsive on iPhone 12 mini

The previous update, iOS 14.2, was a somewhat larger one. It added more than 100 new emoji, incorporated Shazam in the control center, and introduced new audio and AirPlay features, among other things. That update was accompanied by updates to watchOS, tvOS, and others, as well.

Typically, updates with two decimal points in the number are minor bug fix updates, those with just one decimal point are small feature updates, and those with just a whole number (like iOS 14) are annual major releases.

iOS 14.2.1 should be available to all users on supported devices (any carrying the iPhone 12 name) right now.

Continue Reading


iFixit teardown of M1 MacBooks gives us our first glimpse at the M1 up close



As expected, iFixit has done a teardown of two of Apple’s three new M1-based Macs: the MacBook Air and the 2-port, 13-inch MacBook Pro. What they found is somehow both surprising and not: almost nothing has changed in the laptops apart from the inclusion of the M1 chip and directly related changes.

The biggest change is definitely the omission of a fan in the MacBook Air. iFixit notes that given the Intel MacBook Air’s history of overheating in some cases, it speaks volumes about the efficiency of the M1 that so far it seems the Air gets on just fine without that fan now. Also missing: the T2 chip, which we noted in our Mac mini review has been replaced completely by the M1 in all these new Macs.

The 13-inch MacBook Pro is even more similar to its predecessor. The T2 chip is also gone, but the laptop retains the exact same fan and cooling system, with no differences whatsoever. Reviews of the 13-inch MacBook Pro claim that the fan doesn’t spin up as often as it used to, but iFixit concludes here that that’s because of the shift from an Intel chip to the M1, not because of an improved cooling system. The fans on the Intel and M1 Pro are interchangeable.

What’s not interchangeable are a whole bunch of parts in the Air and parts in the Pro. iFixit laments that the similar silicon between the two machines could have presented an opportunity to make repairs easier by making it possible to use parts from one to fix the other, but that seems not to be the case.

And in general, the performance and efficiency gains of the M1 over the prior models are counterbalanced by the fact that user-serviceability and repairability are not moving in a more open direction, because the unified memory architecture of the M1 suggests that Apple isn’t planning to make RAM upgradeable or replaceable any time soon.

iFixit hasn’t given the laptops a repairability score yet, but those two notes suggest the scores wouldn’t be higher than those given to previous models (those machines didn’t have upgradeable RAM, either.)

For more shots of these laptops’ insides and some additional insights, read iFixit’s full teardown post.

Listing image by Samuel Axon

Continue Reading


“We are giddy”—interviewing Apple about its Mac silicon revolution



The graphic representing the Apple M1 chip, as presented by Apple at an event earlier this month.

Some time ago, in an Apple campus building, a group of engineers got together. Isolated from others in the company, they took the guts of old MacBook Air laptops and connected them to their own prototype boards with the goal of building the very first machines that would run macOS on Apple’s own, custom-designed, ARM-based silicon.

To hear Apple’s Craig Federighi tell the story, it sounds a bit like a callback to Steve Wozniak in a Silicon Valley garage so many years ago. And this week, Apple finally took the big step that those engineers were preparing for: the company released the first Macs running on Apple Silicon, beginning a transition of the Mac product line away from Intel’s CPUs, which have been industry-standard for desktop and laptop computers for decades.

In a conversation shortly after the M1 announcement with Apple SVP of Software Engineering Craig Federighi, SVP of Worldwide Marketing Greg Joswiak, and SVP of Hardware Technologies Johny Srouji, we learned that—unsurprisingly—Apple has been planning this change for many, many years.

Ars spoke at length with these execs about the architecture of the first Apple Silicon chip for Macs (the Apple M1). While we had to get in a few inquiries about the edge cases of software support, there was really one big question on our mind: What are the reasons behind Apple’s radical change?

Why? And why now?

We started with that big idea: “Why? And why now?” We got a very Apple response from Federighi:

The Mac is the soul of Apple. I mean, the Mac is what brought many of us into computing. And the Mac is what brought many of us to Apple. And the Mac remains the tool that we all use to do our jobs, to do everything we do here at Apple. And so to have the opportunity… to apply everything we’ve learned to the systems that are at the core of how we live our lives is obviously a long-term ambition and a kind of dream come true.

“We want to create the best products we can,” Srouji added. “We really needed our own custom silicon to deliver truly the best Macs we can deliver.”

Apple began using x86 Intel CPUs in 2006 after it seemed clear that PowerPC (the previous architecture for Mac processors) was reaching the end of the road. For the first several years, those Intel chips were a massive boon for the Mac: they enabled interoperability with Windows and other platforms, making the Mac a much more flexible computer. They allowed Apple to focus more on increasingly popular laptops in addition to desktops. They also made the Mac more popular overall, in parallel with the runaway success of the iPod, and soon after, the iPhone.

And for a long time, Intel’s performance was top-notch. But in recent years, Intel’s CPU roadmap has been less reliable, both in terms of performance gains and consistency. Mac users took notice. But all three of the men we spoke with insisted that wasn’t the driving force behind the change.

“This is about what we could do, right?” said Joswiak. “Not about what anybody else could or couldn’t do.”

“Every company has an agenda,” he continued. “The software company wishes the hardware companies would do this. The hardware companies wish the OS company would do this, but they have competing agendas. And that’s not the case here. We had one agenda.”

When the decision was ultimately made, the circle of people who knew about it was initially quite small. “But those people who knew were walking around smiling from the moment we said we were heading down this path,” Federighi remembered.

Srouji described Apple as being in a special position to make the move successfully: “As you know, we don’t design chips as merchants, as vendors, or generic solutions—which gives the ability to really tightly integrate with the software and the system and the product—exactly what we need.”

Our virtual sitdown included: Greg “Joz” Joswiak (Senior Vice President, Worldwide Marketing), Craig Federighi (Senior Vice President, Software Engineering), and Johny Srouji (Senior Vice President, Hardware Technologies)
Enlarge / Our virtual sitdown included: Greg “Joz” Joswiak (Senior Vice President, Worldwide Marketing), Craig Federighi (Senior Vice President, Software Engineering), and Johny Srouji (Senior Vice President, Hardware Technologies)

Aurich Lawson / Apple

Designing the M1

What Apple needed was a chip that took the lessons learned from years of refining mobile systems-on-a-chip for iPhones, iPads, and other products then added on all sorts of additional functionality in order to address the expanded needs of a laptop or desktop computer.

“During the pre-silicon, when we even designed the architecture or defined the features,” Srouji recalled, “Craig and I sit in the same room and we say, ‘OK, here’s what we want to design. Here are the things that matter.’”

When Apple first announced its plans to launch the first Apple Silicon Mac this year, onlookers speculated that the iPad Pro’s A12X or A12Z chips were a blueprint and that the new Mac chip would be something like an A14X—a beefed-up variant of the chips that shipped in the iPhone 12 this year.

Not exactly so, said Federighi:

The M1 is essentially a superset, if you want to think of it relative to A14. Because as we set out to build a Mac chip, there were many differences from what we otherwise would have had in a corresponding, say, A14X or something.

We had done lots of analysis of Mac application workloads, the kinds of graphic/GPU capabilities that were required to run a typical Mac workload, the kinds of texture formats that were required, support for different kinds of GPU compute and things that were available on the Mac… just even the number of cores, the ability to drive Mac-sized displays, support for virtualization and Thunderbolt.

There are many, many capabilities we engineered into M1 that were requirements for the Mac, but those are all superset capabilities relative to what an app that was compiled for the iPhone would expect.

Srouji expanded on the point:

The foundation of many of the IPs that we have built and that became foundations for M1 to go build on top of it… started over a decade ago. As you may know, we started with our own CPU, then graphics and ISP and Neural Engine.

So we’ve been building these great technologies over a decade, and then several years back, we said, “Now it’s time to use what we call the scalable architecture.” Because we had the foundation of these great IPs, and the architecture is scalable with UMA.

Then we said, “Now it’s time to go build a custom chip for the Mac,” which is M1. It’s not like some iPhone chip that is on steroids. It’s a whole different custom chip, but we do use the foundation of many of these great IPs.

Unified memory architecture

UMA stands for “unified memory architecture.” When potential users look at M1 benchmarks and wonder how it’s possible that a mobile-derived, relatively low-power chip is capable of that kind of performance, Apple points to UMA as a key ingredient for that success.

Federighi claimed that “modern computational or graphics rendering pipelines” have evolved, and they’ve become a “hybrid” of GPU compute, GPU rendering, image signal processing, and more.

UMA essentially means that all the components—a central processor (CPU), a graphics processor (GPU), a neural processor (NPU), an image signal processor (ISP), and so on—share one pool of very fast memory, positioned very close to all of them. This is counter to a common desktop paradigm, of say, dedicating one pool of memory to the CPU and another to the GPU on the other side of the board.

A slide Apple used to present the unified memory architecture of the M1 at an event this year.

A slide Apple used to present the unified memory architecture of the M1 at an event this year.

Samuel Axon

When users run demanding, multifaceted applications, the traditional pipelines may end up losing a lot of time and efficiency moving or copying data around so it can be accessed by all those different processors. Federighi suggested Apple’s success with the M1 is partially due to rejecting this inefficient paradigm at both the hardware and software level:

We not only got the great advantage of just the raw performance of our GPU, but just as important was the fact that with the unified memory architecture, we weren’t moving data constantly back and forth and changing formats that slowed it down. And we got a huge increase in performance.

And so I think workloads in the past where it’s like, come up with the triangles you want to draw, ship them off to the discrete GPU and let it do its thing and never look back—that’s not what a modern computer rendering pipeline looks like today. These things are moving back and forth between many different execution units to accomplish these effects.

That’s not the only optimization. For a few years now, Apple’s Metal graphics API has employed “tile-based deferred rendering,” which the M1’s GPU is designed to take full advantage of. Federighi explained:

Where old-school GPUs would basically operate on the entire frame at once, we operate on tiles that we can move into extremely fast on-chip memory, and then perform a huge sequence of operations with all the different execution units on that tile. It’s incredibly bandwidth-efficient in a way that these discrete GPUs are not. And then you just combine that with the massive width of our pipeline to RAM and the other efficiencies of the chip, and it’s a better architecture.

Continue Reading