What USB adaptor is suitable for a smartwatch?

  • Thread starter Thread starter Shervan360
  • Start date Start date
  • Tags Tags
    Charger
AI Thread Summary
A suitable USB adapter for a smartwatch should output 5V, with the current rating being the maximum the adapter can supply, not what the watch will draw. Using a higher-rated charger (like 15W or 20W) won't damage the watch, as it will only draw the current it needs, typically up to 500mA as per USB specifications. Modern USB chargers are "intelligent" and communicate with devices to adjust current delivery based on their needs. Finding a compatible cable for the smartwatch may be challenging, but as long as the voltage is correct, the adapter will work safely. Overall, any standard USB charger that meets these criteria should be suitable for charging the smartwatch.
Shervan360
Messages
26
Reaction score
4
Hello,

I bought a smartwatch and it does not have a charger. The capacity battery is 230mAh. In the manual, they wrote input parameter should be 5V = 500mA.
If I choose a charger with high output power (15W (5V/3A), (5V/2.4A), 20W, or 65W), will the watch's battery be damaged?
Does the charger with fast charging technology and Power Delivery damage the smartwatch?
Please see the attached image. (The first image is from the smartwatch manual and the last three are Kingstar USB adaptors)
I don't know what USB charger/adaptor to buy.
Thanks in advance
 

Attachments

  • IMG_7939.jpeg
    IMG_7939.jpeg
    46 KB · Views: 106
  • Screenshot 2023-12-24 074740.png
    Screenshot 2023-12-24 074740.png
    28.7 KB · Views: 138
  • Screenshot 2023-12-24 074459.png
    Screenshot 2023-12-24 074459.png
    17.3 KB · Views: 108
  • Screenshot 2023-12-24 083017.png
    Screenshot 2023-12-24 083017.png
    11 KB · Views: 118
Last edited:
Engineering news on Phys.org
As long as the adapter is puts out the normal 5 Volts DC (which USB chargers do), you will be fine -- the watch will draw whatever current it needs up to the limit of the adapter.

Look at it this way: take a look in the fuse box or breaker box where you live. You will find most of the fuses or breakers rated at 15 or 20 Amps. That is the maximum current they will allow before they break the circuit. If you have only a single 60 Watt light bulb turned on, the bulb will only draw the current it needs for 60 Watts.

The current rating on the adapters is the maximum current they CAN supply while keeping their output at 5 Volts -- just like the fuses or circuit breakers, the current rating of a Supply is the maximum available.

For Loads, like your watch or a light bulb, the current rating is the maximum it will ever Need to operate correctly.

The hard part may be finding a cable with a connector to fit the watch!Cheers,
Tom
 
I suppose you can be talking about USB chargers, not any chargers - and as far as I understand these have their own way of doing things.

They are "intelligent" - that is, they start by "talking" to the connected device, asking for the capabilities. If they can't get confirmation "hit me with the high current", they will somehow limit delivered current to just 500 mA, which is a standard USB specification. All higher currents are extensions to the standard, and won't be produced unless specifically allowed by the charged/powered device.
 
  • Like
  • Informative
Likes Tom.G, berkeman and Shervan360
Borek said:
They are "intelligent" - that is, they start by "talking" to the connected device, asking for the capabilities.
You are right, there. I just bought an Apple Watch which has a magnetic charger that holds the watch whilst charging and they say it needs a 40W charger. Then, I notice that most / all new laptops seems to use a USBC socket but specify high charger requirements. That stuff is very clever.
In the distant past, I remember suspecting that hp equipment would communicate with each other because they all seemed to agree with each other to within one sig fig. (This was pre-bluetooth etc)
 
Hey guys. I have a question related to electricity and alternating current. Say an alien fictional society developed electricity, and settled on a standard like 73V AC current at 46 Hz. How would appliances be designed, and what impact would the lower frequency and voltage have on transformers, wiring, TVs, computers, LEDs, motors, and heating, assuming the laws of physics and technology are the same as on Earth?
While I was rolling out a shielded cable, a though came to my mind - what happens to the current flow in the cable if there came a short between the wire and the shield in both ends of the cable? For simplicity, lets assume a 1-wire copper wire wrapped in an aluminum shield. The wire and the shield has the same cross section area. There are insulating material between them, and in both ends there is a short between them. My first thought, the total resistance of the cable would be reduced...
I used to be an HVAC technician. One time I had a service call in which there was no power to the thermostat. The thermostat did not have power because the fuse in the air handler was blown. The fuse in the air handler was blown because there was a low voltage short. The rubber coating on one of the thermostat wires was chewed off by a rodent. The exposed metal in the thermostat wire was touching the metal cabinet of the air handler. This was a low voltage short. This low voltage...

Similar threads

Back
Top