The thing that has been bothering me for a while is that the USB spec allows for software detection of capabilities. You can read the emarker data and see the supported protocols, speeds, voltages, etc.
But there is not standard for usb controllers to present this data to the OS. So it’s stuck in the low level firmware and never passed up. In theory we could have a popup box that tells you that both your computer and other device support higher speeds/more power, but your cable is limiting it.
Apple seems best able to do this since they control the hardware and OS, yet they aren’t doing it either. Users are just left to be confused about why things are slow.
Perhaps someday it will earn the same level of importance as charging; iOS 26 calls out slow chargers on their iPhones, so you can run to the Apple Store and buy a fast one!
They probably have to weigh potential new hardware sales against added complexity. I have counterpoints too but: I believe they try to protect users’ mental models of their ecosystem (which perhaps I appreciate when I don’t notice, and can’t stand when something is uncustomizable). Like there are enough variables they don’t trust us with as it is.
> iOS 26 calls out slow chargers on their iPhones, so you can run to the Apple Store and buy a fast one!
You jest but that notification (it's been a thing on Android for at least 8 years, and on thinkpads for at least 10) has been very helpful to me. Sometimes the negotiation just fails and being told is helpful. Sometimes the charger lies about its specs and once again it's helpful to have a hint, rather than expect everybody to systematically have usb testers on hand.
This one is pretty simple to do. It requests a voltage and then starts pulling current and monitors the voltage as it increases its current draw. If the voltage goes down, alert the user.
With data speed I think it could be a little more complicated. Like OP was saying it would need access to some level of hardware information where it can see which pins are used by the cable. Since the connection 'speed' is still variable even when you DO have a supported cable.
> In theory we could have a popup box that tells you that both your computer and other device support higher speeds/more power, but your cable is limiting it.
I'm pretty sure my old Dell XPS laptop with Windows 10 had pop-ups just like this.
> that's just when plugging in a USB 3 device into a USB 2 port
Dell XPS laptops (and some others) can also warn if the charger isn't providing the full wattage the laptop is rated for. This warning is an option that can be turned off in the BIOS settings.
I usually turn it off because I sometimes intentionally do day trips with a smaller/lighter portable charger which delivers 45w to my laptop which can need up to 65w due to having a discrete GPU. However, 45w is more than sufficient to charge the laptop during normal use on the Balanced power plan with iGPU. I only need more than 45w when gaming with the discrete GPU active.
Just this morning, my old Latitude failed to boot with a “this charger is only giving 20W and that’s not enough to boot this laptop” error. (I was testing a new USB-C charger that’s obviously going back.)
Weirdest part was it was 100% charged, so could have booted with 0 Watts of charger but decided not to boot with 20 Watts more.
Oh, refusing to boot at all is evil. I've never seen that.
Sure, you or I would just unplug the charger and run on battery but bad UX decisions like that generate a support call to me from my 95 yr old mom. It should not only warn and continue to boot, it should use whatever power is on offer to reduce the rate of battery drain.
My wife's work laptop gives this stupid warning anytime any USBC charger is plugged in, other than the Dell brick. So even a dock delivering 100w would get a complaint. The Dell brick offers non-standard charging at 140w, which can't get replaced by standards compliant, smaller chargers.
I wonder if it's possible for a regular machine with two high speed ports to do a cable test by itself. Maybe it can't test all the attributes but could it at least verify speed claims in software?
Apparently the USB driver stack doesn't report the cable's eMarker chip data back to the OS. However benchmarking actual transfer throughput is the ultimate test for data connections (vs charging use cases). Unfortunately, TFA doesn't really go into this aspect of cable testing as the tester seems to only report eMarker data, which pins are connected and copper resistance.
Since a >$1,000 automated lab cable throughput tester is overkill, my thumbnail test for high-speed USB-C data cables is to run a disk speed benchmark to a very fast, well-characterized external NVMe enclosure with a known-fast NVMe drive. I know what the throughput should be based on prior tests with an $80 active 1M Thunderbolt cable made for high-end USB-C docks and confirmed by online benchmark reviews from credible sources.
There would be too many factors involved for a proper test. Many laptop USB controllers would probably not even have the capacity to run two ports at full speed simultaneously.
> But there is not standard for usb controllers to present this data to the OS. So it’s stuck in the low level firmware and never passed up. In theory we could have a popup box that tells you that both your computer and other device support higher speeds/more power, but your cable is limiting it.
There is. I used to use a KVM with USB 2 ports connected to my PC's USB 3 port, to which I connected a monitor with integrated USB 3 hub to drive my keyboard and mouse. Windows would show a popup every time telling me that I should use a faster cable.
There are also popups telling me that my laptop is connected to a "slow" usb-c charger.
That’s quite a simplistic one unfortunately - USB 2 and 3 use different controllers in the PC, which it can indeed detect. The sub-flavours of 3/4 less so.
I've used all manner of archaic usb cables for data transfer when in a pinch and windows has never shown me anything at all. Could it be the external device you were connecting to triggering the windows notification?
Oh, they very much do. But like with everything in technology, they can do fuck all about it, so they resign and maybe complain to you occasionally if you're the designated (in)voluntary tech support person for your family and friends.
Regular people hate technology, both for how magical and how badly broken it is, but they've long learned they're powerless to change it - nobody listens to their complaints, and the whole market is supply-driven, i.e. you get to choose from what vendors graciously put on the market, not from what the space of possible devices.
In general for the 40+ years I’ve been a programmer I have detested the practice of not surfacing diagnostic information to users when technology makes it possible to do so in a clear and unambiguous way.
This is because error messages have historically been bad, unintelligible, un-actionable, and hard to separate from soft errors that don't actually matter.
'Segmentation fault. Core dumped.'
'Non-fatal error detected. Contact support.'
'An error occurred.'
'An illegal operation was performed.'
'Error 92: Insufficient marmalade.'
'Saving this image as a JPG will not preserve the transparency used in the image. Save anyway?'
'Saving as .docx is not recommended because blah-blah-blah never gonna give you up nor let you down.'
I can't blame any normal user from either not understanding nor giving a shit about any of these. If we'd given users actionable information from day 1, we'd be in a very different world. Even just 'Error 852: Couldn't reach the network. Check your connection to the internet.' does help those who haven't turned of their brains entirely yet.
Now imagine if that error said 'Error 11: A memory error occurred. Your program may be faulty or misbehaving. Contact your software vendor." That's miles better than what most things provide.
That one's a good example of why these things are hard. The user could have been running 5 different programs, any one of which caused this error, and MacOS can't point the finger at anyone. Not to mention that the problem could be MacOS itself, or the user being a dunce who misconfigured something. I'm not sure if that error can occur without 3rd party software being involved, but if it can, then that error message might need to be even more vague, helping the user even less. Not to mention it could just be faulty hardware.
A paper manual offering troubleshooting steps for each error would be really helpful. Just 'Error 11. Consult your manual.' and the manual actually telling you what the problem could be is also miles better than what we usually get.
I had a programmer pushing multi-gig packages to a Meta Quest 3; and it was taking around a minute. He didn’t even think that it could be faster because he assumed the Quest or software was slow and didn’t check.
I implored him to try a different cable (after checking cables with the Treedix mentioned in TFA), and the copy went from taking over a minute to about 13s.
Yeah, most programmers are not curious hackers anymore. They are 9-5 white collar workers with hobbies far outside of programming, systems, hardware, etc. It shows very much as soon as you meet one of them. But, like you said, this is true of any industry.
Oh, and pointy jab: these folks are also, in my opinion/experience, the most eager to vibecode shit. Make of that what you will.
"anymore"? Over a decade ago, a coworker had a path for updating some app's files to a database, and it was taking something like 10 minutes on certain test inputs.
Swore blind it couldn't be improved.
By next morning's stand-up, I'd found it was doing something pointless, confirmed with the CTO that the thing it was doing was genuinely pointless and I'd not missed anything surprising, removed the pointless thing, and gotten the 10 minutes down to 200 milliseconds.
I'm not sure if you're right or wrong about the correlation with vibe-coding here, but I will say that co-workers's code was significantly worse than Claude on the one hand, and that on the other I have managed to convince Codex to recompute an Isochrone map of Berlin at 13 fps in a web browser.
I do feel like the industry has taken a nosedive quality wise over covid in particular. Lots of new people only in tech for the money, no deep idea about computers.
But I know stories like yours from a decade past as well. A tale old as time, but compounding in recent years - IMHO.
I think you are right, but I think what I said is also true.
People will notice some things. For example, with USB if they are using it for local backup they might notice, but with a lot of devices they will not. When they do notice, they will feel powerless.
Even if we had a wider choice, they are not well placed to pick products. There is no way they will know about details of things such as USB issues (a cable is slow, the device will not tell you if it is) at the time of purchase.
I think any of us just have to look at how many people ask us for recommendations on basic things like docks and adaptors to see how common this is. On top of that you can’t even trust what’s on the tin sometimes.
This is true of basically everything. Even trivial home maintenance people will just put up with things being broken most of the time over learning how to fix them.
I've lived in this apartment for about a year and a half. It took me until last week to put up lights over the stairs. I've been walking on the stairs in the dark, some times using my phone as a light.
I actually purchased one of these as this article has surfaced before.
It’s well worth the hype, I used it to audit all my cables (both for home and work) and it’s amazing how many thick and unwieldy cables are actually terrible for data.
For example I purchased a pair of B&W Px8 S2 noise cancelling headphones, which boast a DAC if you connect via USB-C directly, the cable it came with though was thick but only rated for USB 2.0 speeds. These headphones cost more than AirPods Max, which are themselves considered overpriced, and include comforts like nappa leather; so shipping with a chunky cable that doesn’t even carry decent data feels like a bizarre oversight. Apple’s own USB-C cables manage the same power delivery at less than half the thickness with a woven shell. You’d assume a premium product would at least match that.
Honourable mention to the USB-C cables that ship with Dell Ultrasharp monitors (both pre-USB4 and post). Those support basically everything except Thunderbolt 4 despite being unmarked.
> so shipping with a chunky cable that doesn’t even carry decent data feels like a bizarre oversight.
USB 2.0 can support up to 480 Mbps. It’s more than fast enough for any audio stream you can send to a DAC.
Your headphones don’t need USB 3.0 5 Gbps speeds. USB 3 requires extra wires with different properties that need to be controlled more tightly, which can impact cable flexibility. If your headphones used USB 3 when they didn’t need it that would be one more thing to break and more failure modes for the cable.
A USB 2 cable with fewer conductors was the right choice for this product. The fact that you only got miffed about it when plugging the cable into a tester, not from actually using the product or cable, is good evidence that a USB 3 cable wasn’t needed.
Nobody said the headphones needed USB 3. The point is that the cable is physically thick and rigid (like something you'd expect to carry serious data) but doesn't. Meanwhile Apple ships a thinner, more flexible cable that supports the same USB 2.0 speeds and equivalent power delivery. The cable B&W chose is worse ergonomically for no functional benefit. That's the kind of mismatch the Treedix exposes.
I started buying Belkin TB5 cables which are around $50 a pop. They can easily power a laptop at full load and can stream video at any reasonable resolution/framerate I might need. I've yet to find a need for an NVMe faster than 20 GBps nevermind finding USB4 enclosures, or that the cable supports up to 80. They're also not nearly as chunky as the Dell cables, which are good, but seem to have very rigid shielding.
I keep a few converters for older devices and servers that don't have (m)any C ports, but as far as a consumer "forever cable" goes, TB5 feels close. Certainly the cable's bandwidth is beyond what most people need, unless you're editing 8k video or continually shuffling hundreds of GBs between external disks.
The point isn’t really about audio bandwidth; it’s about the cable being strangely overbuilt for what it actually does.
It’s rigid and thick, like a Thunderbolt 3 cable, yet only supports USB 2.0 speeds and fast charging for a device that doesn’t need fast charging.
Compare that to Apple’s iPhone USB-C cable which is thin, flexible, and supports the same features.
That matters because someone might grab that cable assuming it’s a “better cable”: it came with a £629 product, it’s thick and feels serious, so surely it’s capable. But it isn’t. And there’s nothing marked on it to tell you otherwise.
The whole system ends up relying on presumption, which is exactly the problem the device in the article is solving.
> The point isn’t really about audio bandwidth; it’s about the cable being strangely overbuilt for what it actually does.
The purpose of the heavy construction is to make it durable, not to carry 5 Gbps data streams to your headphones.
Unlike most USB peripherals like your printer and keyboard that get plugged in and then don’t move around, headphone cables go to your head and move around constantly. They can get pinched in drawers or snagged on corners.
Apple's woven USB-C cable gets dragged around with iPhones, iPads and laptops daily and manages durability at half the thickness. Durability doesn't require rigidity... in fact for a headphone cable, rigidity is the opposite of what you want. Stiff cables tug on the headphones and transmit mechanical noise.
Maybe Apple's changed their cables recently, but the fragility is the reason I avoid Apple cables.
Especially in headphones. The number of times those broke during a bike ride or run was way to high for me to keep wasting money on them knowing full well they weren't going to last more than a few months just like every other Apple headphone I've ever had.
It’s common to add weights to headphones to make them feel premium which is bizarre since actually premium headphones tend to try very hard to reduce weight as the weight makes them more uncomfortable.
I don’t know how to fix the market especially when consumers keep rewarding these practices, and I think the effectiveness of TikTok style influencer marketing will make it worse.
I don’t think that’s what’s happening here. B&W actually reduced the weight on the Px8 S2 compared to the original, and the headphones themselves are genuinely lightweight for what they are. The cable isn’t thick to “feel premium” (it feels kinda bad); it’s thick because it’s rated for 65W+ power delivery that the headphones don’t need.
The problem is the opposite of what you’re describing, it’s not a cynical design choice, it’s a lazy one. They probably just purchased a cable for capabilities irrelevant to the product and the result is worse ergonomics and misleading physical cues about what the cable can actually do.
“I don’t think..” Ok, you’ve made a number of assumptions and we don’t share the same priors so I’m unable to follow you to your conclusion.
I think you are underestimating the importance of perceived premium combined with the pressures of cost accounting, but I do think that is pretty normal for ‘audiophiles’ which is their target market.
Which assumptions? The weight reduction on the S2 is documented and the cable’s 65W rating is what the tester confirmed.
If the argument is that B&W deliberately chose a thick cable to seem premium, it doesn’t square with them actively slimming down the headphones. B&W are primarily a speaker company, their USB-C product range is basically just a few headphones and earbuds.
More likely they just sourced a generic cable that happened to support high wattage and didn’t think about the mismatch.
Either way, we’re deep in the weeds on B&W’s cable procurement now. The root point is that USB-C is a mess. You can’t tell what a cable supports by looking at it, and even premium manufacturers are shipping cables that don’t do what you’d reasonably expect.
That’s exactly the problem the Treedix from the article solves.
My point on weight was that the market for that it is common, which is probably a stronger statement than needed. I should have made the weaker argument and said the market exists which only needs one example. The company Beats can serve as that example, this company sells the majority of premium headphones but I don’t actually know what percentage have weights placed in them. I am assuming a non trivial percentage.
You are using circular reasoning in your logic, you assume the premise is true and from there you derive your evidence.
I would contend that someone thought about it and decided to go with the cheaper option because they could get away with it. I would consider my assumption to have more grounding given my experience with manufacturing and cost accounting.
You’ve gone from “companies add weight to feel premium” to “they went with the cheaper option because they could get away with it.” Those are opposite explanations. But either way, the cable doesn’t do what its physical presence suggests, nothing on it tells you otherwise, and that’s the entire point of the device in the article.
My position is entirely consistent, it is cheaper to signal premium quality than actually deliver it. The point I am making is that there is immense comercial pressure to do this is a highly competitive market when selling to consumers who don’t know better.
My example of weights is that the steel weighs are cheaper than the alternative of using heavier drivers, by adding weight they are signaling premium without delivering it. Similarly with the USB cable, consumers assume such cables are thick because of thicker wires and better shielding, it’s cheaper to make a thick cable without those those features, once again signaling premium without actually providing it.
That's a more coherent version of your argument, but it's still speculative. You're attributing a deliberate strategy to what is more easily explained by indifference. B&W make about four products with USB-C cables. This isn't a company with a cable strategy, cynical or otherwise.
4th times the charm. You’ve provided no evidence for indifference. My point remains, given industry standards indifference would be highly unusual and not at all a safe assumption.
The vast majority of high volume consumer manufacturers use cost accounting practices which would absolutely be tracking and attributing the usb cable costs and the whole point of that accounting practice is to constantly be thinking about minimizing costs of even the smallest inputs, all the way down to the individual screws used. Yes, they’re thinking about how to save 1/100ths of a cent from each screw.
The reason it is thick is because it supports 65W charging. Apple did the same with the USB-C cables that shipped with the pre-MagSafe MacBooks. It was a thicker cable that supported 100w charging but was only USB 2.0.
I think that's being a bit uncharitable to B&W specifically; they're one of the few headphone companies where the engineering does back up the price. The cable is the odd one out.
The headphones have equivalent performance whether a USB 2 cable is connected, or a USB 3 cable is connected. The headphones themselves are not USB 3 devices; the addition of USB 3 cabling instead of USB 2 cabling would change absolutely nothing about how they work.
So, no: I wouldn't expect the cable for a pair of headphones (of any price) to support USB 3. That represents extra complexity (literally more wires inside) that is totally irrelevant for the product the cable was sold with. (The cables included with >$1k iPhones don't support USB 3, either.)
Meanwhile: Fast charging. All correctly-made USB C cables support at least 3 amps worth of 20 volts, or 60 Watts. This isn't an added-cost feature; it's just what the bare minimum no-emarker-inside specification requires. A 25-cent USB C-to-C cable from Temu either supports 60W of USB PD, or it is broken and defiant of USB-IF's specifications.
---
Now, of course: The cable could be thinner and more flexible and do these same things. That'd probably be preferred, even: Traditional analog headphones often used very deliberately thin cables with interesting construction (like using Litz wire to reduce the amount of internal plastic insulation) to improve the user's freedom of movement, and help prevent mechanical noise from the cables dragging across clothes and such from being telegraphed to the user's ears.
Using practical cabling was something that headphone makers strived to be good at doing. I'm a little bit annoyed to learn that a once-prestigious company like B&W is shipping cables with headphones that are the antithesis of what practical headphone cables should be.
---
But yeah, both USB C cables and the ports on devices could be better marked so we know WTF they do, to limit the amount of presumption required in the real world. So that a person can tell -- at a glance! -- what charging modes a device accepts or provides, or whether it supports video, or whether it is USB 2 or USB 3, or [...].
Prior to USB C, someone familiar with the tech could look at a device or a cable and generally succeed at visually discerning its function, but that's broadly gone with USB C. What we have instead is just an oblong hole that looks like all of the other oblong holes do.
After complaining about this occasionally since the appearance of USB C a decade or so ago, I've come to realize that most people just don't care about this -- at all. Not even a little bit. Even though these things get used by common people every day, the details are completely out of the scope of their thought processes.
It doesn't have to be this way, but it's not going to change: Unmarked ports are connected together with unmarked cables and thus unknown common capabilities are just how we roll.
No. CD audio is 1,4 mbit. Even increasing the temporal and spatial resolution beyond that, which is audiophile nonsense, will never even approach USB 2 speeds.
I wasn't surprised to learn that when Linus Tech Tips released those new usb-c cables, that they all sold out almost instantly. They put their entire reputation on the line to claim (and label) the exact capabilities of their usb cables. Isn't that all we really want?
Yeah I don't see any problems with their stuff. It ain't exactly cheap, but a large part of that is their work in making sure you aren't just getting some random 2 dollar trash.
Yup, that's the sort of thing that's typically missing from cable testers. I have a USB cable that normally works fine, but introduces errors when doing full blast USB 2.0 bulk transfers. I keep it around just in case I ever come across a tester that can show me this in hard numbers.
What I'm looking for is a differential signal tester, where you can breakout any arbitrary cable or traces and test the properties of the wire with different frequencies. It should be able to measure interesting properties such as resistance, capacitance, inductance, phase/length difference, wire length, etc.
One of these devices for approximately $100 would sell all day long.
As someone who really doesn't care about learning the details, and just want one USB-C cable that does it ALL to put in my backpack what should I buy ?
You don’t really want that. A thunderbolt cable is both stiff and expensive. They only really make sense to leave attached to the back of a monitor or dock.
What would work better is a flexible 100w+ usb3 cable. You can’t do thunderbolt on it but it’s a tiny fraction of the cost and does everything you’d actually need on the go.
So much this. I have a few different categories of "known good" USB-C cables because one type doesn't fit all my use cases. Sometimes the trait I need is >100w PD charging at 1M. Sometimes I need 80 Gbps dual 4k video at 3M. Other times I need 40 Gbps .5M to a portable NVMe enclosure. USB-C cables I regularly rely on range from $5 to $100 and weight/size varies >3x.
And in my tiny 'go bike bag' for day trips I need one 2M cable that's thin, coils into a tight ball and weighs nothing yet will charge up to 45w and reliably xfer data at up to 5Gbps (USB 3.1) for quick uploads with optional USB-A and Micro-USB adapters at either end (because I still know people with Micro-USB (though it obviously drops to USB2 speeds)).
At my workplace someone always orders the what they perceive to be the "best" cables. They aren't thunderbolt, they are just oversized with thick braiding. They are all so stiff and heavy you can barely handle a phone while charging without the cable pulling itself out.
Love this! I got a USB C multimeter and used it to test the output of two dozen chargers. Wanted to see if they supplied the voltage that was advertised. Funny enough, AOHI was the only brand whose chargers actually increased their voltage as my current draw went up. It was like the engineers knew about the resistance in the wire and decided to compensate by upping the voltage slightly.
As an alternative, you could get a stand-alone USB-C power meter which can be used with any cable. That way, when the cable breaks, you don't have to buy a new power meter. Here is an example of one such product (though I've never used this model): https://www.amazon.com/Adapter-Voltage-Current-Extension-Con...
And to be precise, a nice, high quality thunderbolt cable from a reputable manufacturer like Apple or OWC. Protect the cable as it will have been expensive, but it will work very well.
One thing to realize is that especially for high resolution video cables these cheap testers can't really deliver. The way to test them is a eye diagram (see: https://incompliancemag.com/eye-diagram-part2/ ) and testers with that capsbility cost upwards of 10.000 Eurodollars.
No. What it can affect though is the bandwidth of the cable, meaning e.g. for HDMI cables, they might not support higher resolutions or framerates. If it's on the border you might see random disconnects or screen blanks.
The quality degrading is not something you will see, as it's a digital protocol.
"Audiophile grade" HDMI cables are likely to just be a Shenzhen bargain-bin special with some fancy looking sheathing and connectors. I would trust them less than an Amazon Basics cable.
Indeed. If I want super high quality cables, I get them from Blue Jeans Cables, who tell you exactly what Belsen or Can are cable stock and what connectors, as well as the assembly methodology.
Correct. But especially if you're using long cables a cable with more "headroom" in the eye diagram will perform more reliable than one that is just at the edge of breakup.
For home use that doesn't matter usually, but I for example run events where I need the cable to work also after 10 people stepped on it and then this can become a significant thing.
These two statements aren't mutually exclusive. The link is looking at the analog signal through an oscilloscope. The person you replied to is pointing out that after decoding and applying error correction, you can still end up with the same digital signal output. So the eye diagram charts are useful for detecting the quality of the cable, but as long as the quality is past a certain threshold, it does not matter.
No. What I am saying is that it is hard to test the quality of a 8K 240Hz 4444 video cable without having a device that can send and receive this or even higher.
If you send bits across a line fast enough you're grtting into the territory of RF electronics, with wrong connector or conductor geometry you will get echos on the line and all kind of signal loss. A good digital protocol should keep this at bay with error correction and similar mechanisms, but if you want to know what the good cable is on a better than binary scale of works/does not, you need to look at these things.
Well the thing is better doesn't mean better quality here. Better means you can use a longer cable or abuse the cable for longer till it dies.
This is a big part of what makes any pro gear expensive: reliability. If you just connect your home hifi to your speakers in an acoustically untreated space, you could also just use a bunch of steel wire coathangers and get an indistinguishable result. Even a el-cheapo store brand music shop cable will do the trick for years if you don't habitually change your setup four times a week (most people don't).
But if you need reliability and predictability in a studio or live context giving a damn about cable quality is mandatory since a broken cable in the wrong place can ruin your day and reputation. But it is an absolute myth that they will affect the sound in any meaningful way.
Exeption: guitar cables. The capacitance of guitar cables can shift the resonance frequency of the pickup up or down leading to audibly different results. But that id no magic either, you could just take a low capacitance cable and add in arbitrary capacitor for 10 cents as needed.
Brilliant little device. I will be picking one up ASAP!. Didn't know that lying cables were a thing but I have a ton of charge only cables?!
I speculate USB B wasn't included because there are only really two types, 2.0 (regular size) and 3.0 (has an obvious extension on the connector). There also don't tend to be power-only A-B cables because they are usually found on printers, Arduino s, ... And not for charging devices.
Fun fact: A Xiaomi fast charge cable (with orange plugs) has an extra contact on the A end to support USB C PD out of a USB A charger.
I wanted to have a model which tells me the modes which are supported and which is actually selected for a reasonable price and which I can order at a reasonable trader. this model seems to do the trick
Similarly: Is there a USB-C power delivery adapter to force directionality? I needed to siphon off power from small batteries into a larger pack (that could supply more power out than the small packs) in a power outage. I absolutely could not force my larger power station to accept a charge and it kept pushing power back the wrong direction despite which ends of the cable I plugged in first.
That cable has one power input (that is only an input), and two outputs (that are only outputs), and a brainbox in the middle to direct the circus.
If we label the connectors as A, B, and C, then it works like this: A charges B and/or C, and other charging directions are no-op.
The less-complex way is to use a USB A to C cable, if that's appropriate. With these, the A side is always the source and the C side is always the sink.
---
And yeah, it's annoying. I got a cheap lithium car jump starter several years ago with some neat power bank features (like 60W USB PD in/out, on one port). So I plugged it into my phone with USB C at my desk, and discovered that they'd charge eachother seemingly randomly. While changing nothing, I'd look over and sometimes the jump starter would charge the phone, and sometimes the phone would be charging the jump starter. The conglomeration formed a heater, with more steps.
But there is not standard for usb controllers to present this data to the OS. So it’s stuck in the low level firmware and never passed up. In theory we could have a popup box that tells you that both your computer and other device support higher speeds/more power, but your cable is limiting it.
Apple seems best able to do this since they control the hardware and OS, yet they aren’t doing it either. Users are just left to be confused about why things are slow.
They probably have to weigh potential new hardware sales against added complexity. I have counterpoints too but: I believe they try to protect users’ mental models of their ecosystem (which perhaps I appreciate when I don’t notice, and can’t stand when something is uncustomizable). Like there are enough variables they don’t trust us with as it is.
You jest but that notification (it's been a thing on Android for at least 8 years, and on thinkpads for at least 10) has been very helpful to me. Sometimes the negotiation just fails and being told is helpful. Sometimes the charger lies about its specs and once again it's helpful to have a hint, rather than expect everybody to systematically have usb testers on hand.
With data speed I think it could be a little more complicated. Like OP was saying it would need access to some level of hardware information where it can see which pins are used by the cable. Since the connection 'speed' is still variable even when you DO have a supported cable.
I'm pretty sure my old Dell XPS laptop with Windows 10 had pop-ups just like this.
"This device can run faster" or something.
Dell XPS laptops (and some others) can also warn if the charger isn't providing the full wattage the laptop is rated for. This warning is an option that can be turned off in the BIOS settings.
I usually turn it off because I sometimes intentionally do day trips with a smaller/lighter portable charger which delivers 45w to my laptop which can need up to 65w due to having a discrete GPU. However, 45w is more than sufficient to charge the laptop during normal use on the Balanced power plan with iGPU. I only need more than 45w when gaming with the discrete GPU active.
Weirdest part was it was 100% charged, so could have booted with 0 Watts of charger but decided not to boot with 20 Watts more.
Sure, you or I would just unplug the charger and run on battery but bad UX decisions like that generate a support call to me from my 95 yr old mom. It should not only warn and continue to boot, it should use whatever power is on offer to reduce the rate of battery drain.
Since a >$1,000 automated lab cable throughput tester is overkill, my thumbnail test for high-speed USB-C data cables is to run a disk speed benchmark to a very fast, well-characterized external NVMe enclosure with a known-fast NVMe drive. I know what the throughput should be based on prior tests with an $80 active 1M Thunderbolt cable made for high-end USB-C docks and confirmed by online benchmark reviews from credible sources.
There is. I used to use a KVM with USB 2 ports connected to my PC's USB 3 port, to which I connected a monitor with integrated USB 3 hub to drive my keyboard and mouse. Windows would show a popup every time telling me that I should use a faster cable.
There are also popups telling me that my laptop is connected to a "slow" usb-c charger.
I don’t know if they check that via USB protocol, or if they are measuring the actual power draw on the USB port.
In order to use the device, I had to connect it via an externally powered USB hub.
Regular people hate technology, both for how magical and how badly broken it is, but they've long learned they're powerless to change it - nobody listens to their complaints, and the whole market is supply-driven, i.e. you get to choose from what vendors graciously put on the market, not from what the space of possible devices.
They hate having to go through people that get them upset, in order to use their kit.
Not just tech (although it’s more prevalent). People who are “handy” can also be that way (but, for some reason, techies tend to be more abrasive).
I’ve learned the utility of being patient, and not showing the exasperation that is often boiling inside of me.
In general for the 40+ years I’ve been a programmer I have detested the practice of not surfacing diagnostic information to users when technology makes it possible to do so in a clear and unambiguous way.
"What did the error message say"
"I don't know."
'Segmentation fault. Core dumped.'
'Non-fatal error detected. Contact support.'
'An error occurred.'
'An illegal operation was performed.'
'Error 92: Insufficient marmalade.'
'Saving this image as a JPG will not preserve the transparency used in the image. Save anyway?'
'Saving as .docx is not recommended because blah-blah-blah never gonna give you up nor let you down.'
I can't blame any normal user from either not understanding nor giving a shit about any of these. If we'd given users actionable information from day 1, we'd be in a very different world. Even just 'Error 852: Couldn't reach the network. Check your connection to the internet.' does help those who haven't turned of their brains entirely yet.
"I don't understand, it says 'System Error Type 11', and no matter how many times I type 11, nothing happens!"
That one's a good example of why these things are hard. The user could have been running 5 different programs, any one of which caused this error, and MacOS can't point the finger at anyone. Not to mention that the problem could be MacOS itself, or the user being a dunce who misconfigured something. I'm not sure if that error can occur without 3rd party software being involved, but if it can, then that error message might need to be even more vague, helping the user even less. Not to mention it could just be faulty hardware.
A paper manual offering troubleshooting steps for each error would be really helpful. Just 'Error 11. Consult your manual.' and the manual actually telling you what the problem could be is also miles better than what we usually get.
I implored him to try a different cable (after checking cables with the Treedix mentioned in TFA), and the copy went from taking over a minute to about 13s.
Its not just normal people confused.
Oh, and pointy jab: these folks are also, in my opinion/experience, the most eager to vibecode shit. Make of that what you will.
Swore blind it couldn't be improved.
By next morning's stand-up, I'd found it was doing something pointless, confirmed with the CTO that the thing it was doing was genuinely pointless and I'd not missed anything surprising, removed the pointless thing, and gotten the 10 minutes down to 200 milliseconds.
I'm not sure if you're right or wrong about the correlation with vibe-coding here, but I will say that co-workers's code was significantly worse than Claude on the one hand, and that on the other I have managed to convince Codex to recompute an Isochrone map of Berlin at 13 fps in a web browser.
But I know stories like yours from a decade past as well. A tale old as time, but compounding in recent years - IMHO.
People will notice some things. For example, with USB if they are using it for local backup they might notice, but with a lot of devices they will not. When they do notice, they will feel powerless.
Even if we had a wider choice, they are not well placed to pick products. There is no way they will know about details of things such as USB issues (a cable is slow, the device will not tell you if it is) at the time of purchase.
I'm an electrician.
It’s well worth the hype, I used it to audit all my cables (both for home and work) and it’s amazing how many thick and unwieldy cables are actually terrible for data.
For example I purchased a pair of B&W Px8 S2 noise cancelling headphones, which boast a DAC if you connect via USB-C directly, the cable it came with though was thick but only rated for USB 2.0 speeds. These headphones cost more than AirPods Max, which are themselves considered overpriced, and include comforts like nappa leather; so shipping with a chunky cable that doesn’t even carry decent data feels like a bizarre oversight. Apple’s own USB-C cables manage the same power delivery at less than half the thickness with a woven shell. You’d assume a premium product would at least match that.
Honourable mention to the USB-C cables that ship with Dell Ultrasharp monitors (both pre-USB4 and post). Those support basically everything except Thunderbolt 4 despite being unmarked.
USB 2.0 can support up to 480 Mbps. It’s more than fast enough for any audio stream you can send to a DAC.
Your headphones don’t need USB 3.0 5 Gbps speeds. USB 3 requires extra wires with different properties that need to be controlled more tightly, which can impact cable flexibility. If your headphones used USB 3 when they didn’t need it that would be one more thing to break and more failure modes for the cable.
A USB 2 cable with fewer conductors was the right choice for this product. The fact that you only got miffed about it when plugging the cable into a tester, not from actually using the product or cable, is good evidence that a USB 3 cable wasn’t needed.
Apple’s iPhone cables are not known for their durability. They serve a mostly stationary purpose, unlike headphones you wear on your head.
I keep a few converters for older devices and servers that don't have (m)any C ports, but as far as a consumer "forever cable" goes, TB5 feels close. Certainly the cable's bandwidth is beyond what most people need, unless you're editing 8k video or continually shuffling hundreds of GBs between external disks.
It alleviates the anxiety of knowing what cable does what.
I use Apples Thunderbolt 4 or USB-C cables exclusively: if its white its for charging and low data, if its black its for high data.
I’ve been doing this for a few years, but its really costly as those Apple Thunderbolt cables are crazy expensive.
It’s rigid and thick, like a Thunderbolt 3 cable, yet only supports USB 2.0 speeds and fast charging for a device that doesn’t need fast charging.
Compare that to Apple’s iPhone USB-C cable which is thin, flexible, and supports the same features.
That matters because someone might grab that cable assuming it’s a “better cable”: it came with a £629 product, it’s thick and feels serious, so surely it’s capable. But it isn’t. And there’s nothing marked on it to tell you otherwise.
The whole system ends up relying on presumption, which is exactly the problem the device in the article is solving.
The purpose of the heavy construction is to make it durable, not to carry 5 Gbps data streams to your headphones.
Unlike most USB peripherals like your printer and keyboard that get plugged in and then don’t move around, headphone cables go to your head and move around constantly. They can get pinched in drawers or snagged on corners.
Hence the more durable construction.
Apple’s USB iPhone cables wearing out prematurely is so common it’s a meme.
Maybe Apple's changed their cables recently, but the fragility is the reason I avoid Apple cables.
Especially in headphones. The number of times those broke during a bike ride or run was way to high for me to keep wasting money on them knowing full well they weren't going to last more than a few months just like every other Apple headphone I've ever had.
https://www.techgearlab.com/topics/electronics/best-usb-c-ca...
I don’t know how to fix the market especially when consumers keep rewarding these practices, and I think the effectiveness of TikTok style influencer marketing will make it worse.
The problem is the opposite of what you’re describing, it’s not a cynical design choice, it’s a lazy one. They probably just purchased a cable for capabilities irrelevant to the product and the result is worse ergonomics and misleading physical cues about what the cable can actually do.
I think you are underestimating the importance of perceived premium combined with the pressures of cost accounting, but I do think that is pretty normal for ‘audiophiles’ which is their target market.
If the argument is that B&W deliberately chose a thick cable to seem premium, it doesn’t square with them actively slimming down the headphones. B&W are primarily a speaker company, their USB-C product range is basically just a few headphones and earbuds.
More likely they just sourced a generic cable that happened to support high wattage and didn’t think about the mismatch.
Either way, we’re deep in the weeds on B&W’s cable procurement now. The root point is that USB-C is a mess. You can’t tell what a cable supports by looking at it, and even premium manufacturers are shipping cables that don’t do what you’d reasonably expect.
That’s exactly the problem the Treedix from the article solves.
You are using circular reasoning in your logic, you assume the premise is true and from there you derive your evidence.
I would contend that someone thought about it and decided to go with the cheaper option because they could get away with it. I would consider my assumption to have more grounding given my experience with manufacturing and cost accounting.
My example of weights is that the steel weighs are cheaper than the alternative of using heavier drivers, by adding weight they are signaling premium without delivering it. Similarly with the USB cable, consumers assume such cables are thick because of thicker wires and better shielding, it’s cheaper to make a thick cable without those those features, once again signaling premium without actually providing it.
The vast majority of high volume consumer manufacturers use cost accounting practices which would absolutely be tracking and attributing the usb cable costs and the whole point of that accounting practice is to constantly be thinking about minimizing costs of even the smallest inputs, all the way down to the individual screws used. Yes, they’re thinking about how to save 1/100ths of a cent from each screw.
Or, why Apple manages the same in half the footprint?
Or, why someone would expect that a cable that came with a pair of headphones actually charges things at over 65w?
So, no: I wouldn't expect the cable for a pair of headphones (of any price) to support USB 3. That represents extra complexity (literally more wires inside) that is totally irrelevant for the product the cable was sold with. (The cables included with >$1k iPhones don't support USB 3, either.)
Meanwhile: Fast charging. All correctly-made USB C cables support at least 3 amps worth of 20 volts, or 60 Watts. This isn't an added-cost feature; it's just what the bare minimum no-emarker-inside specification requires. A 25-cent USB C-to-C cable from Temu either supports 60W of USB PD, or it is broken and defiant of USB-IF's specifications.
---
Now, of course: The cable could be thinner and more flexible and do these same things. That'd probably be preferred, even: Traditional analog headphones often used very deliberately thin cables with interesting construction (like using Litz wire to reduce the amount of internal plastic insulation) to improve the user's freedom of movement, and help prevent mechanical noise from the cables dragging across clothes and such from being telegraphed to the user's ears.
Using practical cabling was something that headphone makers strived to be good at doing. I'm a little bit annoyed to learn that a once-prestigious company like B&W is shipping cables with headphones that are the antithesis of what practical headphone cables should be.
---
But yeah, both USB C cables and the ports on devices could be better marked so we know WTF they do, to limit the amount of presumption required in the real world. So that a person can tell -- at a glance! -- what charging modes a device accepts or provides, or whether it supports video, or whether it is USB 2 or USB 3, or [...].
Prior to USB C, someone familiar with the tech could look at a device or a cable and generally succeed at visually discerning its function, but that's broadly gone with USB C. What we have instead is just an oblong hole that looks like all of the other oblong holes do.
After complaining about this occasionally since the appearance of USB C a decade or so ago, I've come to realize that most people just don't care about this -- at all. Not even a little bit. Even though these things get used by common people every day, the details are completely out of the scope of their thought processes.
It doesn't have to be this way, but it's not going to change: Unmarked ports are connected together with unmarked cables and thus unknown common capabilities are just how we roll.
Of course they are advertising their own new USB cable, but as someone who didn't know much about USB cables I find it quite interesting.
One of these devices for approximately $100 would sell all day long.
> Access Denied
> Sorry, you do not currently have the necessary permissions to access this site, or this site may not be available in your region.
Are they geoblocking the USA from even viewing their site for some reason?
It seems to be a more comprehensive "Make sure the lines go where they are supposed to" tester. Looks pretty good.
But the devices that test things like transmission speed, are a lot more expensive.
I think that many of the issues that this device tests, can be mitigated by simply buying cables from reputable sources.
Even in my reputable cables, there are a couple with suspicious continuity issues. I wonder if this could find them.
You could probably build a data transfer tester using an FPGA and some signal processing.
Don't forget the speeds at which modern serial interfaces go. Being able to look at the data, at that speed, requires some serious kit.
I didn't know there were cable testers like this, thank you.
What would work better is a flexible 100w+ usb3 cable. You can’t do thunderbolt on it but it’s a tiny fraction of the cost and does everything you’d actually need on the go.
If you actually do want it, this is the do everything cable https://www.apple.com/au/xc/product/MW5H3ZA/A
And in my tiny 'go bike bag' for day trips I need one 2M cable that's thin, coils into a tight ball and weighs nothing yet will charge up to 45w and reliably xfer data at up to 5Gbps (USB 3.1) for quick uploads with optional USB-A and Micro-USB adapters at either end (because I still know people with Micro-USB (though it obviously drops to USB2 speeds)).
https://iaohi.com/products/aohi-the-future-adonis-usb4-2-0-2...
https://news.ycombinator.com/item?id=47561827
(these are what I would buy from a sea of cables, not the cheapest, but far from the most expensive)
I imagine at that length and speed, signal integrity becomes difficult.
The quality degrading is not something you will see, as it's a digital protocol.
"Audiophile grade" HDMI cables are likely to just be a Shenzhen bargain-bin special with some fancy looking sheathing and connectors. I would trust them less than an Amazon Basics cable.
For home use that doesn't matter usually, but I for example run events where I need the cable to work also after 10 people stepped on it and then this can become a significant thing.
Not in terms of quality, but reliability.
If you send bits across a line fast enough you're grtting into the territory of RF electronics, with wrong connector or conductor geometry you will get echos on the line and all kind of signal loss. A good digital protocol should keep this at bay with error correction and similar mechanisms, but if you want to know what the good cable is on a better than binary scale of works/does not, you need to look at these things.
This is a big part of what makes any pro gear expensive: reliability. If you just connect your home hifi to your speakers in an acoustically untreated space, you could also just use a bunch of steel wire coathangers and get an indistinguishable result. Even a el-cheapo store brand music shop cable will do the trick for years if you don't habitually change your setup four times a week (most people don't).
But if you need reliability and predictability in a studio or live context giving a damn about cable quality is mandatory since a broken cable in the wrong place can ruin your day and reputation. But it is an absolute myth that they will affect the sound in any meaningful way.
Exeption: guitar cables. The capacitance of guitar cables can shift the resonance frequency of the pickup up or down leading to audibly different results. But that id no magic either, you could just take a low capacitance cable and add in arbitrary capacitor for 10 cents as needed.
I speculate USB B wasn't included because there are only really two types, 2.0 (regular size) and 3.0 (has an obvious extension on the connector). There also don't tend to be power-only A-B cables because they are usually found on printers, Arduino s, ... And not for charging devices.
Fun fact: A Xiaomi fast charge cable (with orange plugs) has an extra contact on the A end to support USB C PD out of a USB A charger.
That cable has one power input (that is only an input), and two outputs (that are only outputs), and a brainbox in the middle to direct the circus.
If we label the connectors as A, B, and C, then it works like this: A charges B and/or C, and other charging directions are no-op.
The less-complex way is to use a USB A to C cable, if that's appropriate. With these, the A side is always the source and the C side is always the sink.
---
And yeah, it's annoying. I got a cheap lithium car jump starter several years ago with some neat power bank features (like 60W USB PD in/out, on one port). So I plugged it into my phone with USB C at my desk, and discovered that they'd charge eachother seemingly randomly. While changing nothing, I'd look over and sometimes the jump starter would charge the phone, and sometimes the phone would be charging the jump starter. The conglomeration formed a heater, with more steps.
(Back and forth with the same poop, forever.)
--- (I remember. The poop.)