Author Topic: Panasonic NKY467B2 36V 15AH 540Wh. Ebike Li ion upgrade, burning my father'ass?  (Read 14969 times)

0 Members and 1 Guest are viewing this topic.

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
I want to upgrade my father's chinesium ebike from stupid AGM to Li ion battery. It's a crappy bike but it was a present...
After the chinesium charger melted the 3x12V (already replaced once) batteries, I throw those stupid heavy (my mother don't even want to touch the bike because it's too heavy) AGMs with the charger out of the window.

So I got an used ebike li ion battery from evilbay for a very reasonable price, plus I got my self another used ebike lion charger again stupid cheap. Cheap IS a must, I am not gonna to spend money on a sick horse.  :horse:

The battery I got it's a Panasonic NKY467B2 36V 15AH 540Wh, and the charger it's 42V 2A.




In the description the charger gauge reported 4/5 state of charge, so far so good.

Today it arrived and I started to play with it. Battery voltage was 22V, freaking 22V but the gauge always shows 4 of 5 led on... so I can't trust those leds. Oh well...


 
I started the rescue mission.



First I open the battery... very well done. It's a 900€ (new!) battery, so I was expecting something good inside.
The package is a potting fest inside a nice plastic bag with am air vent (goretex?). there are two main connectors: one for the bike/big charger and one for a small charger unit.

Here some pics



smaller charger connector



I took the battery from 22V to about 30V with 100mA charging current. Then I moved with about 1C to 36V... now it has just reached 36V. Of course I check the single cells voltage and they are always pretty close the others. yes. Who knows if the battery is still good... meh.



It's a 10S3P configuration BTW, I am almost sure.

Now I have to understand how that BMS works and rev-eng it...
Do you guys knows what these letters means on the bike/big charger connector?



+ and -, I know it
C: the charger port
D: ? (Drive?)
S: ?

Let's see if my father's ass will burn on that upgraded bike. Safety first...  :-DD

Any help is appreciate. More to come.
« Last Edit: October 09, 2018, 09:25:04 am by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline John Heath

  • Frequent Contributor
  • **
  • Posts: 444
  • Country: ca
  • 2B or not 2B
Nice test bench. BMS stands for battery management system. BMS is your friend as it will shut the entire battery pack down if any of the lithium batteries ventures below 3 volts discharging and stop charging at 4 volts. This is good as discharging a lithium below 3 volts is throwing the dice. You can get away with it a few times but once too often and the lithium is toast. I would leave the BMS alone. The bike will run fine on 38 volts so why change it , a bird in the hand...

On another note if the battery pack was stupid cheap then chances are there is a bad lithium which will limit the bike range when the BMS shuts it down. With BMS you are only as good as the weakest lithium battery. The way to tell is cycle the battery pack through a charge and discharge. Towards the end of the discharge measure the individual lithium batteries. If most say 3.5 volts but 1 says 3.1 then you found the bad guy. With a little luck they will all say the same voltage + - 200 m volts so you are good to go. :-+

I was going to leave at this but I have to know. What is the square box in the lower right corner of the test bench with the round circle on it?? A compass ?   
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
I was going to leave at this but I have to know. What is the square box in the lower right corner of the test bench with the round circle on it?? A compass ?

Do you mean the Nespesso coffee box?

Thanks John to stay with me in this madness. I charged 1C up the battery  to 40V and the 10 cells were all at about 4.00V so cool  :clap:.
No temperature rising whatsoever, it looks good.

I don't like that BMS for two reasons:

1) The state of charge was showing 4/5 full with a 22V battery voltage. It can't be right.
2) I think there are MOSFET



https://www.st.com/resource/en/datasheet/stp260n6f6.pdf

to enable the + and the charging port... and I think that's why I didn't have 40V on the external bike connector between + and -. This means I need to rev eng also before to attach the battery charger.

Best case the BMS is stuck in error and power cycle with a charged battery will do.

Now I am logging the self discharge voltage of the battery with that BMS attached... will see.


« Last Edit: October 09, 2018, 09:27:11 am by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
Ok, instead of rev eng I will remove completely that BMS and go with these cheap cheap chinesium toys

BMS, I hope 20A cont./30A peak are enough:



eBay auction: #263901952562


Battery Gauge:



eBay auction: #162766012576


and done deal.

PS: Don't you love the technical description of those cheap toys?
I think I will connect the BMS like this:



cristal clear, thank God there are just + and - in a battery...
How many time they put a - instead of a +? :horse:  |O



EDIT: This makes more sense:

« Last Edit: October 09, 2018, 11:59:06 am by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
Keep the original BMS! It will be way better than a replacement. The state of charge usually is determined by current going in & out and it will be tuned for the particular cells used. If you do one discharge / charge cycle the state of charge will show the right value. Unlike other chemistries you can not use the voltage for the state of charge in Li-ion batteries. The original BMS probably shows a wrong value because the cells self-discharged.
« Last Edit: October 09, 2018, 09:28:34 am by nctnico »
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
If you do one discharge / charge cycle the state of charge will show the right value.

Hi nico, thanks for jumping in. The problem I don't know how to properly charge or discharge using that BMS, I don't have the V Bat voltage accessible (on both charge or discharge positive cables) on those connectors because the BMS is cutting them off somehow. I could buy the 200€ charger for it and give it a go, reverse eng it and ship it back... meh...
Not sure if it is worth it.

Surely it would be better to keep it, but that crap bike will not be used for critical missions: 10 minutes travel in the town and come back with fresh bread. Who cares if the SoC is not precise.
What I need is a way to see if the charger has done his job: 42V? Yes then full, dad you can disconnect the charger.

On top of that a proper reverse eng would require a lift of that board to see what is on the other side... so I will destroy the potting anyway... and hoping to understand something.
14€ and job done. Surely chinesium quality job, but it's a chinesium bike.  :-\  ::)
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
If you charge to 40V then it will be OK. A Li-ion charger will use a current limit until it reaches a certain voltage. Form there the voltage is kept constant. The charge ends when the current drops below a certain threshold. Say 200mA for this kind of battery.
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
According to my tests, here the 10€ used li ion ebay ebike charger I have (from my memories, I have a written notes at home):

22V<VBat<40V: 2A. (I don't like the 2A @<33V, meh)
40V<VBat<42V: reducing almost linearly the 2A
42V: 0A

I think we are in business.

Only missing pieces in the puzzle are the thermal analysis and the hope that Max 20A cont., 30A Peak 36V will be enough for this sick horse:



which is very similar to mine... oh according to the www those are 250W bike, so 250/36= 7A  :-+

 :popcorn:

 
« Last Edit: October 09, 2018, 12:49:07 pm by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
Back home now, the self discharging test looks promising



Moreover here the cheap ass li ion ebike charger tests results:

Vout [V]Iout [A]
20.52.248
41.752.248
41.801.988
41.851.707
41.881.528
41.941.2
41.971.021
420.831
42.0600.5
42.10.261
42.12490.1
42.1400

meh.. I can live with it for a 14Ah battery 36V.
« Last Edit: October 09, 2018, 05:06:46 pm by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
Unlike other chemistries you can not use the voltage for the state of charge in Li-ion batteries.

Completely the opposite.

On li-ion (except maybe LFP), the voltage is way better estimate for SoC than on almost any other chemistry. On NiCd, NiMH, lead acid, SoC estimate based on voltage is almost useless, because the discharge curve is almost a flat line: the voltage difference between 80% and 20% may be just 5%, or even less.

This is simply because li-ion has the open-circuit voltage range running from about 4.2V (100%) to about 3.4V (0%), and on most li-ion chemistry variants, there is no "flat" part on the line. While the curve isn't linear, it isn't a "straight line with only a sudden drop" either.

On li-ion, the voltage difference between 80% and 20% is typically around 15%, making voltage measurement feasible way to approximate the SoC. It's often accurate enough, especially with a non-linear lookup table. Especially older consumer stuff simply used voltage. Typically "good enough", expect +/-20% accuracy when somewhat calibrated for the particular cell type. This is way better than a poorly designed coulomb counter which can totally fail and easily produce 100% error.

But depends on the exact li-ion chemistry. You can look at the actual curves collected, for example, here:
https://lygte-info.dk/review/batteries2012/Common18650CurvesAll%20UK.html

LFP is an exception, with its almost flat discharge curve.
 

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
Unlike other chemistries you can not use the voltage for the state of charge in Li-ion batteries.
Completely the opposite.

On li-ion (except maybe LFP), the voltage is way better estimate for SoC than on almost any other chemistry. On NiCd, NiMH, lead acid, SoC estimate based on voltage is almost useless, because the discharge curve is almost a flat line: the voltage difference between 80% and 20% may be just 5%, or even less.
I think you got this completely the other way around. The charge/discharge curve on most Li-ion cells is a flat curve for a big part of the cycle and the temperature has a much bigger influence on the cell voltage than the charge level. IOW: the cell voltage is completely useless to determine the amount of charge (SOC). This is the reason why quality BMSses use a charge gauge which is calibrated for the specific chemistry of the cells reaching accuracies of 1%. One of my customers makes Li-ion battery so I know a thing or two about Li-ion cells and BMSses.
« Last Edit: October 10, 2018, 04:23:52 pm by nctnico »
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
Please check the link I gave you with the actual curves. They have a lot! You can easily see they are not flat. My measurements give similar curves. This is again one funny discussion, it's always a strange feeling when someone comes from around the corner and tells something you have been doing on a daily basis for years just fine is "impossible" or "does not work"  |O. But let's be more specific here, I hope this info helps anyone:

I have designed and built battery measurement & cycle testing equipment and performed probably millions of test cycles on at least 30 different cells of LFP, NMC, NCA, LCO and LMO chemistries. I have designed and small-scale commercialized a few BMS systems as well, and built two concept EV conversion battery systems (20 kWh and 40kWh) to test said systems on the road. One of the cars was in daily driving for years I think, giving a lot of data to look at at varying real-life conditions. I was doing some of this research for some time at a university (including curve testing, internal resistance testing, self discharge testing, capacity fade testing, and DCR rise testing); I had some very good chances to communicate with the chief chemist of a commercial li-ion manufacturer there. These kinds of contacts are really valuable when you need to deal with the massive flow of true and false information on the 'net - even in academic research.

Then I went freelancer to design another piece of cycle testing equipment. We did quite a few measurements for companies designing large li-ion equipment. The problem with the li-ion cells is: even if you are a customer buying in tens of millions, you are not big enough to get real data on the products. So you need to test. And to do that, you outsource it to a test lab. A test lab, then, has a $1000000 setup of MACCORs. Or, you can use some HobbyKing  style battery testers. We tried a PowerLab - they are utter jokes and toys (as expected). So I designed some sanely working redistributive (energy-saving) "accurate enough" (to 0.5% typically) equipment which can automatically cycle, perform different pulsing for DCR measurements, inject specific AC current waveforms, measure AC resistance, log capacity, voltage, temperature... With configurable channels, we could run 20 different cells at +/- 20A each, or run fewer tests with paralleled channels. We had some 100A discharge tests going on some 6Ah cells for over a thousand of cycles. We used fridges and heatbeds to prove things the customers didn't want tested, but we knew how the chemistry works and needed to prove the point. And so on...

Anyways, to the results, and back to the question of do you need coulomb counting or is voltage measurement enough:

With LFP (lithium iron phosphate), I find coulomb counting is almost absolutely necessary, since the discharge curve is indeed flat. You are probably thinking about this specific chemistry. Which has become quite a niche IMHO.  To be fair, some other chemistries are somewhat as flat as well - I have seen some fairly flat NMC (nickel - manganese - cobalt oxide) cells as well. Not as flat as typical LFP, but I'd expect these are difficult. But these are in minority.

In expensive, large pack EVs, I use coulomb counting as well, because 1) the relative cost isn't prohibitive; 2) and because users often find they need an exact, steadily dropping "km/miles range left" indication, and become panicked if the range jumps near the end; and, last but not least importantly: 3) these systems tend to be fully charged after almost each cycle, or almost every day, giving a reliable reset point for the integrator. This is really important! No one builds an everyday BMS with a 0.1% or 0.01% precision current sensing. These exist in said $1M MACCORs. +/-2% is reality, and this means it drifts below the accuracy of a simple voltage based system in just a few cycles if not reset. And, reseting is always based on voltage; the trick is it happens on well known conditions, such as 4.20V at C/20 charge rate.

But in simple, small gadgets, I always tend to lean towards voltage-based estimation; sometimes I'm so lazy I just do it as a linear approximation - not a single customer has complained about the nonlinearity of even these solutions! It runs faster when it's starting to get empty, and wiggles around with the load a bit. Depends on the exact specifics how usable it is, but your blanket statement it's not usable is clearly untrue. I find it usable more often than not.

Agreed, using a chemistry-specific lookup table provides much better results, and is utterly trivial to implement - download a curve of your cell of choice from lygte-info.dk who probably already tested it and you are done! Yes, the readout wiggless with changing load - but that's intuitive for the customer. It can be a good feature: what good it makes to have a steady 10% SoC value on screen, then the product suddenly dies on you when you press the pedal or push a button, with the load peak hitting LVC limit behing the scenes? With the voltage based estimate, you start seeing it wiggle alarmingly towards 0% momentarily, and you know it'll die anytime soon with the next peak; it doens't take a li-ion chemist to understand this UI behavior. Very intuitive, especially on a graphical bar scale. It's "good enough" out of the box, without complex algorithms or UI design.

Voltage based approximation has one very good trait: it's stateless - and it automatically tracks the remaining actual capacity with absolutely no algorithm! The accuracy is what it is (I say typically around +/-20% when people want to have a simple single number), but it works very robustly in cases where natural reset points for the integrator do not happen. It's an absolute classic to see a completely invalid battery level integration on a coulomb counting based system just after about ten cycles, this has always happened with laptop battery management, for example, and people have complained about it for ages. What's the value in the "+/-1% best case accuracy", if the worst case accuracy is +/- 50%, and the algorithm cannot tell you whether to trust it or not?

This being said, really well designed coulomb counters can, of course, fall back to the voltage mode, do some kalman filtering like correction all the time to the integrator value, and apply temperature and current compensations to get, say, +/-10% accuracy in voltage mode and +/-1% accuracy in the integration mode. While I have seen discussions of such algorithms, and given quite some "brain masturbation" on this idea while walking in the forest hunting (edible) mushrooms, I have never seen such system implemented in practice. Maybe some magical IC is already getting there? I don't know - I don't trust them after seeing so many total failures.

It's worth mentioning that coulomb counting the exact SoC, even if accurate, serves only one single purpose: the user experience. This info is typically not used for battery management purposes.

I did quite some work as a BMS failure analyst. Totally unplanned, it just happened that the failing systems were everywhere. Typically very complex, overengineered. Although the most typical failure mode is that they imbalance the pack and overdischarge a cell (or several cells), it's also very typical they implement broken-by-design coulomb counting algorithms.

Use what tools you need to use - but if you are unsure, and if you are not really up to the task, choose the simplest, and naturally most robust way to do that, if ever possible! A four-five-LED indication of the approximate voltage is really useful in real life. Something like that has always been used in approximate battery gauges, but it only works for alkaline cells - and li-ion, surprisingly!
« Last Edit: October 10, 2018, 05:51:01 pm by Siwastaja »
 
The following users thanked this post: Wolfram, MT

Online SiliconWizard

  • Super Contributor
  • ***
  • Posts: 15030
  • Country: fr
In my experience, estimating a charge level on a Li-ion/LiPo battery based on battery voltage only is often pretty inaccurate and has many pitfalls. For one, it depends so much on the actual current draw profile that it's hard to use in many applications in which the current draw is pulsed and not at all constant, with periods of low load (maybe in the µA range) and peaks in the tens or hundreds of mA for instance. For a constant current load, that's already much more predictable, but a lot of applications are not a constant current load.

I now tend to use ICs that do a good job at this task ("battery fuel gauge") while using nicely developed algorithms based on voltage, current and accumulated energy, such as the MAX1704x series.
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
I now tend to use ICs that do a good job at this task ("battery fuel gauge") while using nicely developed algorithms based on voltage, current and accumulated energy, such as the MAX1704x series.

Thanks for pointing this out - if you read the datasheet, you'll see they point out the problems with coulomb counting.

Yes. This IS a voltage-based device! They don't measure the current. This is exactly what I have been talking about.

I wonder how you didn't notice you didn't connect any current sense leads anywhere while using this device? How do you think it measures current?

They show an example error analysis in one particular test case with -7%/+3% error over the SoC. This is fairly typical for a curve-compensated, temperature-compensated algorithm. It's probably way more accurate in worst case conditions (probably around +/- 15%) over the classical ones that use current sensing and coulomb counting (with underdeveloped algorithms) and can show any random value.

The best thing? I'm 99% sure their touted magical "sophisticated battery model" which "simulates the internal dynamics of a Li+ battery" is something simple and trivial  ;). The 1k | 1u RC filter they suggest rules out most "AC trickery" as well. Since they don't measure the current, they can't compensate for it (directly).

Just filtering the shit out of the signal (for example, with cumulative moving average on the MCU, with a time constant in range of half a minute), you have a nice smooth number like in coulomb-counted systems. So you'd never know it's "just" voltage based!

OTOH, I would assume that for such 7% error, you would need to:
"To  achieve  optimum  performance,  the  MAX17043/MAX17044  must  be  programmed  with  configuration  data  custom  to  the  application. Contact the factory for details."

This would include the cell curve lookup table, I guess.

It might be easier to just upload the graph from lygte-info.dk to your MCU and do it yourself. Depends on how keen you are on developing your own I guess. The obvious plus side would be the BOM and complexity savings, since this IC doesn't provide any "analog" feature you need, if your MCU has an ADC, this is just some secret code running on a separate device; a way to license the algorithm would make so much more sense.
« Last Edit: October 10, 2018, 05:45:30 pm by Siwastaja »
 

Online SiliconWizard

  • Super Contributor
  • ***
  • Posts: 15030
  • Country: fr
I mentioned the MAX series as an example and those are the ICs I currently use for this task (I like the elegance of MAXIM's approach and find them reasonably accurate). But I've also used other similar ICs that were based on coulomb counting only or a mix of current measurement and coulomb counting, and those were not that bad actually. Certainly much better than just predicting the SOC based on just the loaded cell voltage on a proportional scale as I've seen done a few times. Using a cell voltage measurement in addition to the estimated consumed charge helped getting more accurate predictions AFAIR, but that was pretty much calibrated on a given battery model.

If you think about it, even though MAXIM doesn't disclose their algorithm, current is still indirectly used IMO. For a given battery model, variations of the cell voltage on a short term will give you a good indication of the drawn current if you have enough resolution. They may not *directly* rely on current estimation, but I think it still gets its influence indirectly in their estimation.

As they state and as I remember from someone working at SAFT, the open circuit voltage of a Li-ion battery is actually a good indicator of its SOC. But in a real, always-on application, the OCV can't be directly measured (usually), so this is an indirect estimation based on the loaded cell voltage as far as I've gotten it, so I guess they most likely use "short"-term variations as well as the average value on a longer term.
 

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
To clarify: the battery gauge chips (like the ones from TI) obviously do more than just measuring charge (current going in/out) like measuring end-of-charge/discharge voltage, keep track of aging but the secret sauce is never revealed.

If you measure only the voltage then from what I've seen and read you'll have problems with varying loads and temperatures so if you are going to make a voltage based charge monitor you'll need tables with voltage versus temperature and current to determine the amount of charge going in/out. That seems more complex (or less accurate) than measuring the charge going in & out and adjust that with reading the voltage at points in the charge/discharge curve where it varies a lot.

@Siwastaja: it is difficult to asses where someone's knowledge comes from. You have obviously done your homework.
« Last Edit: October 10, 2018, 07:33:00 pm by nctnico »
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline jbb

  • Super Contributor
  • ***
  • Posts: 1194
  • Country: nz
I have recently been looking at the Texas Instruments Impedance Track goodies.

They basically do a voltage lookup, where the LUT is chemistry-specific.

They estimate the open circuit voltage using measured voltage,  current and temperature (several more lookups). To account for gariation and ageing, some parameters are learnt in situ; most chips include a Coulomb counter to help with that.

Currently having a go with a BQ40Z50-R2, which does 1 to 4 series cells, and includes protection, gauging and (low current) balancing. Optionally generates commands for a smart charger.

I hear a lot along the lines of “don’t use a BMS, they kill your batteries.” But you at least need voltage protection on a per-cell basis and current protection for the pack overall, so I think it really should be “don’t use a crap BMS.”
 

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
“don’t use a crap BMS.”

Well I just bought a cheap ass 10S Li Ion chinesium BMS as stated before. How can I test it if it is crap or not?
I imagine to use:
- 10 R ladder to simulate the 10 Li Ion cells voltages
- DC load to simulate battery when charging or load for battery in normal use
- one or two PSU to simulate charger or battery

By adding some R in parallel to the 10 R ladder I can simulate a low cell and see how the BMS behave... and so on.

Any other suggestions?

BTW: Thank you all, I am learning a LOT.
« Last Edit: October 11, 2018, 09:34:45 am by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
so I think it really should be “don’t use a crap BMS.”

You don't need to look too far back to see where this is coming from. Not even a decade ago, the idea of using a "good BMS" was mostly theoretical for most people: "crap BMS" was the standard. It was impossible to know for a product design engineer - and it still is quite difficult; even for li-ion experts.

Out of commercial li-ion BMS chips (mostly flooded to the market by TI), most were broken-by-design in a way or another. There still are many traps, especially by TI. I have evaluated several totally broken-by-design li-ion ICs by TI. Probably some are very good. But how can I trust them? Looking at TI's li-ion product datasheets, it's utterly clear these parts are designed by inexperienced engineers in a hurry, not using any kind of high-reliability design practices, to keep up with the management's desire to have the greatest lithium ion management IC portfolio (50 new chips out every year, even though the problem field is static); not in co-operation with battery chemistry experts, reliability experts or safety experts.

Out of commercial li-ion BMS modules, most were outright dangerous, or at least would kill the battery. Of course, they are using said IC's, or some are home-brew solutions.

Then you have the more expensive hi-tech li-ion management ICs which come with a ton of certifications and paperwork. I looked at one ADI part for example, with separate redundant analog comparator-based backup system for the cell voltage measurements, against ADC/MCU failures. Sounds great! Then, looking at the typical application in the appnote, both shutdown signal paths were brought out to one microcontroller which "decides" about the system shutdown. Every idiot on the planet can laugh at this ridiculousness. Yet, it's a really well certified high-reliability product for automotive.

Now, what comes to li-ion safety, people tend to make these completely wrong assumptions:
* Li-ion is super unsafe; it instantly blows up or catches fire if you overcharge it the slightest
* BMS is the magic sauce which prevents all those fires which would happen otherwise.

In reality:
* A poor quality no-name Alibaba li-ion cell may be fairly dangerous; otherwise, a modern, typical li-ion cells implement cell and chemisty level protections such as shutdown separators (that shut down the ionic transfer before the onset of thermal runaway), PTC resettable fuses, CIDs (interrupters based on cell pressure). I have tried my best to try to blow up a modern Samsung/Sony/Panasonic/LG cell. I haven't succeeded. I have applied 30V, 10A for 8 hours to a 4.2V cell. The cell becomes a self-regulating hysteretic heater that swiches on and off at about 120 degC. The plastic wrapping changes its tint. Worst case, I got a tiny leak of electrolyte out. These are BTW all tests that are specified by the manufacturer; they guarantee their cells pass such tests. External heat over the thermal runaway onset temperature - around 150 degC - would be the best bet, since the modern shutdown separators seem to work so well that even a nail penetration is not setting these things on fire anymore.

Warning: this is not to say you should abuse the cells in any way. They can catch fire because the inherent chemistry is still very volatile - it's just got safety layers built around it -, and abusing them will of course increase the risk as it puts more burden on the safety features that are not "normally" needed. It's just that it doesn't tend to usually happen, because the safety features are well designed.

* The cell-level, balancing BMS, when working properly, may extend the usable pack lifetime slightly. A non-working, destructive BMS often does not cause safety problems, because - see the previous point. The cells are usually OK with the abuse a faulty BMS gives them. So, a faulty BMS just destroys the pack, but very seldom in a dramatic or dangerous way - this is 100% thanks to the li-ion chemists and engineers!


So, all this is why I rolled my own BMS, not using any BMS IC, for EVs and energy storage (scalable, for large packs, up to 250s, up to around 100-200kWh; something similar to the modules available back then, such as Elithion, just a simplified, minimalistic design.) I did semi-commercialize it (producing a few full units, selling at very low price to selected customers). Even though I tried to address all issues seen in failing BMS's, I still don't have complete trust in my own, either.

Why don't I trust my design? Because I have seen very experienced professionals fail to provide the reliability. I have balancing! It can get stuck on, and although that doesn't cause fire - because I did thermal analysis for stuck-on balancing - it could still overdischarge a cell. I have a timeout feature (some TI products don't - they kill your battery automatically if your I2C communication fails stuck once in your product lifetime!) But still.

BMS design is non-trivial. The first issue you face is defining what you need to do, the basic functionality and specifications. This is hard due to information overflow. The focus is easily lost to difficult-to-implement but unimportant features, such as:
* high balancing currents
* redistributive balancing
* complex algorithms not based on actual battery science, but the "gut feeling" instead. For example, at lot of effort has gone into AC measurements and "state-of-health analysis" bullshit instead of just implementing reliable LVC and HVC basics.

Then, when it comes to implementation:
* A BMS needs permanent connections to dozens of cell taps, possibly over a decade. Powering any electronics continuously, so that it's guaranteed to work within tight specs for a decade is non-trivial.
* Leakage currents need to be kept to minimal levels, even in corner cases. Any kind of latch-up of increased leakage - such as an MCU or an ASIC FSM exiting sleep and staying awake, or getting stuck to measuring loop - is a catastrophe which automatically kills the battery.
* If high balancing currents are involved in a dissipative way, the power dissipation analysis cannot be done by the "typical" basis, assuming short duty cycle. Balancing resistor can get stuck on for several reasons. I remember at least one reported conversion EV fire that was likely to be caused by overheating balancing resistor.

To make the point: compare the MTBF for a "MCU or FSM gets stuck in a wrong state" event for general consumer (or even industrial or automotive!) electronics, and a BMS.

The general device:
* Runs maybe a few hours a day
* Has a typical lifetime of probably five years; after that, no one's interested.
* Resets every now and while, when power cycling
* May get stuck without causing much problems: the user just resets it and we are good to go again!

Think of any gadget, even well designed. Imagine that every time you need to boot it for any reason, it would die instead. That's the level of reliability we need to think about when designing a cell-level BMS, especially with balancing.

On a li-ion BMS, a full reset cannot be done. It's permantenly powered for a decade; often in a difficult (read automotive) environment. What's worst, such a failure event almost guarantees the self-destruction of the pack! If any part of the IC / MCU gets stuck, a reset cannot be done, power cycling cannot be done, it's all hardwired inside the enclosure. It looks dead, is nonresponding, and you just wait it to kill the cells with it.

The MTBF for the similar event should be at least 4-5 orders of magnitude longer than for general consumer electronics. And because no typical BMS designer - not even at TI, they are making low-cost product series - has access to some super high-reliability NASA space technology, what do you think? It's all based on lowest cost commercial off-the-shelf processes. Especially at TI.

Which is why the only way I could imagine increasing the reliability was to simplify, reduce part count and complexity. But this isn't going to make the 4-5 order of magnitude difference required. So, I'm not too happy in my design. One cell module has actually failed on the field (but, luckily, didn't kill the cell). I suspect ESD during manufacturing or calibration.

When I was hired in the university I talked above, they actually had this super expensive conversion EV with a super over-engineered BMS (with redistributive balancing and everything). The total BOM count for the 80-cell system was over 5000 components, tens of meters of wire, around 300 connectors... And the problem was, the BMS was in some peculiar "state", it had been for half a year at that point, didn't let the car boot, didn't enable the charger. Now, when they finally let us start dismantling the car, about 30% of the cells were already completely dead, discharged to 0V. Now, the only task why the BMS actually exists is to isolate the battery pack when any cell hits LVC or HVC, completely preventing overdischarge. This BMS failed exactly its primary purpose. To the designers, it clearly wasn't primary.

We got to see another similar EV case a year later, with a very different kind of BMS, and it had the exact same story: the BMS consisted of about ten 8-cell modules (so again around 80 cells total), and out of these ten modules, two were latched up, in a way that they had killed all 8 connected cells through the balancing taps. So, 16 cells were completely dead, 64 cells completely OK.

--

Really, the essence of a cell-level BMS is 90% cargo cult. You just design one in, now you have a BMS! And you don't need to think about it. Convenient, huh!

It's likely to be some random product from TI's massive lineup, most likely broken-by-design. I have chosen TI li-ion management part twice in my life in my own designs (I don't understand why, I usually learn from the mistakes of others), and regretted it twice, and redesigned it twice. It has wrong or non-optimal setpoints, it somehow does let the cells overdischarge, then does "preconditioning" at an order of magnitude higher current (I have seen C/20 in a TI product) than what's considered safe and instructed by battery manufacturers (typically C/100).

It claims to have "overvoltage protection", but when you look at the block diagram, you'll notice it's connected in the wrong place in hurry by the designers, so it has no chance of protecting anything. It's next to impossible to find all these traps beforehand.

Oh well, I had a pack charge to 4.63V/cell on a very simplistic prototype (with no secondary, redundant protection) by a TI part which was fully functional and could have just shut down the MOSFET it was actively driving "on" - from charger input to the battery -  despite the internally nonconnected "overvoltage" signal screaming out of their lungs. But it worked perfectly! My lazily done linear voltage-based battery gauge showed 127% and everybody was happy, because the extra charge was there, and really extended the runtime ;D. No fire. Thanks Samsung for a great product. No thanks go to TI.

The Typical TI BMS mostly works by luck, it might kill a small percentage of products after some years, but not too many - and no one even thinks about the cause. They think: "oh, the batteries are just unreliable, thank God we have a BMS, without it we would be seeing higher failure rates I'm sure!" Then, in some cases, the BMS causes some theoretically dangerous error, such as lets the cells overcharge - but, thanks to the robustness of the modern cells, nothing dramatic happens. So, everybody's happy, products are reliable enough for most people, and the BMS checkbox is ticked!

But you at least need voltage protection on a per-cell basis and

This is a very interesting myth I see recurring - if I had a dollar every time....

It has some basis on it, but it certainly isn't a general and "hard" rule like people think.

Not using cell-level voltage measurements for cutoffs is not only limited by cheap Aliexpress specials.

Since you seem to know this, could you explain to me why Robert Bosch does not need "per-cell basis" voltage protection? I mean, they are fairly reputable I think?

Could you explain the huge number of li-ion charge management ICs, that are supposed to be used with two cells connected in series, without center tap monitoring, as a single 7.2V nominal cell?

There is no debate about connecting 2 cells in series without cell-level cutoffs. That's the absolute industry standard practice, has been since day 1. The only debate about this is by hobbyists, on forums.

On over 2 cells, or large packs, things start get more complicated, like always, getting us to the "it depends" territory. But Robert Bosch isn't the only reputable manufacturer who has no issues going up to 6s without cell-level anything. There are some industry design traditions: laptops always have cell-level monitoring and balancing (and "killed by BMS" packs were fairly typical at one point about a decade ago - I have dissected several) - power tools often won't (and I haven't seen a single incident of a imbalanced pack, or a cell at 0V).

current protection for the pack overall,

Safety-wise, the most reliable current protection is a passive fuse, properly sized (not massively oversized). Don't ever forget this back up in case your MOSFET switches fail short. Remember to look at fuse DC ratings and DC breaking currents.

Sorry for getting so verbose. Hope this all helps someone.
 
The following users thanked this post: Zucca, MT

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
Well I just bought a cheap ass 10S Li Ion chinesium BMS as stated before. How can I test it if it is crap or not?

There is no way to test that reliably, easily. Really useful testing takes a lot of time and resources.

OTOH, quick, basic testing gives you a initial go/no-go quickly. For this,

Quote
Any other suggestions?

I suggest using in actual environment, in the bike, and carrying a multimeter with you to do some verifications. Especially look at the signals when charging. You can always set the charger CV setpoint to slightly higher than normal, during controlled testing, to see if the BMS can turn the charger off on HVC event. Then, bring the CV setting back so now you have a layer of safety - if the BMS fails, the charger still limits the voltage like it would in a non-BMS system. The only way the BMS can now fuck this up is by stuck-on balancer causing a massive imbalance.

The effect of LVC, you can verify while driving the pack empty. At some point, it just suddenly stops. If it's getting sluggish instead, take your multimeter and verify the cells are still in balance, even on the bottom. With modern high-quality cells, capacity matching is good enough that cells tend to be fairly well in balance on both top and bottom. And if the cells are in balance on bottom, the risk of low-voltage damaging them by driving is small, even if the cell-level LVC fails to work - it will get so sluggish that you just stop. So better watch the balance when near empty, that will give you the warning sign of which cells to watch more closely.


But at some point, it's just OK to follow the cargo cult and not worry too much. If the cells are of proper quality, the risks are fairly small.

The Chinese BMS design is probably not any worse, and is likely to be much better than a typical over-engineered Western amazingly novel super BMS.
« Last Edit: October 11, 2018, 08:34:51 am by Siwastaja »
 
The following users thanked this post: Zucca

Offline ZuccaTopic starter

  • Supporter
  • ****
  • Posts: 4468
  • Country: it
  • EE meid in Itali
Siwastaja

Once I started I could not stop reading all of it, I was addicted to it. I thank you so much.
You not only inspired me to go deeper in my chinese dead horse bike project, but also now I fell like one day I will/can/should develop a Li Ion energy storage pack for my future home.

There are so many questions and comments I want to write now, but I prefer to digest it first, do some experiments and report back.

I will ask Dave to create a Battery section in the forum. Probably better not to have it, most of the know how should be in the section "Renewable Energy"

Quote
Renewable Energy: Solar, wind, thermal, nuclear, energy storage, Electric Vehicles etc.

Ops I should have posted this there....
« Last Edit: October 11, 2018, 09:22:31 am by zucca »
Can't know what you don't love. St. Augustine
Can't love what you don't know. Zucca
 
The following users thanked this post: Siwastaja

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
@Siwastaja: I'm wondering: is any of your BMS designs used in large production runs of battery packs? From what I've seen the TI chips are pretty common to use as an analog front-end for a microcontroller. If the chips from TI where as bad as you describe nobody would use them.
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
@Siwastaja: I'm wondering: is any of your BMS designs used in large production runs of battery packs? From what I've seen the TI chips are pretty common to use as an analog front-end for a microcontroller. If the chips from TI where as bad as you describe nobody would use them.

Oh, the classic "million flies" argument  :)! OK, thanks for your interest really, let me explain my point of view a bit more.

I never dared to commercialize the "full home brew" BMS in large scale, as I explained above, so I kept it small scale. In total, I had 500 and later another 500 cell modules made. I still have some. I want to be careful and avoid making the same mistakes others did. And while my design reduced the complexity and BOM cost by about 3-10x compared to the competitors, this doesn't prove ultimate reliability.

So, if my manifesto says that the existing li-ion products are too unreliable, I would need to prove A) that they really are; and B) that mine are considerably better, and proving or disproving reliability takes a lot of time and resources! I would also be fighting against industry giants, with hard-to-really-prove claims. Not proving every claim may be OK on forum discussions (no one expects it from themselves, only from the others who disagree), but for really professional work on safety and reliability, this marketing speech right here wouldn't cut it.

And I'm not actually a reliablity-nut at all. I'm mostly just fine with the reliability of the TI chip. I'm also fine with an occasional lithium ion fire happening, if the thing is sorted out properly, root cause investigated and fixed.

What I'm not fine with is all the hypocrisy and lack of information around these solutions - let alone the absolute classic: the false sense of security for those who are more concerned about the safety than I am. What I'm also not fine with is the BMS cargo cult, which manifests itself very well in the way how BMS manufacturers need to document in big red letters that yes, you indeed need to connect your BMS shutdown signals to actually shut something down. (For example, see the very first words of Elithion manual: http://lithiumate.elithion.com/php/index.php ) This usage pattern which really exists is not too far from using a wooden BMS unit with bamboo wires that looks real. Even on these forums, I have heard this argument multiple times:

"BMS is absolutely critical for safety".
Then, when questioned, they go on:
"Geez, just put in some BMS chip and you are done!"

They put quite a lot of trust on that random, unnamed chip, and quite a lot of trust on the application engineer deploying it properly, for such "safety critical" thing!

BTW, AFAIK, and correct me if I'm wrong, but most of the "TI chips" and alike do not have any kind of safety qualifications done (they wouldn't probably pass), and if the product is in relevant group so that you need the safety stuff done, you as a system integrator take the full risk, and you need to understand the big picture completely, all the small details included.

Only the most expensive chips very few people end up using have the paperwork done (which doesn't prove much, as explained earlier).

. . .

So I went on to make case-by-case custom solutions where I need li-ion batteries as a part of something larger, because I like large problems, often done in "good enough" manner. I always try to prefer a single-cell solution or max 2s-3s (without cell-level management) and step up the voltage, if possible, and "manage" the cell by the MCU that sits there anyway. If necessary, I employ completely redundant analog backups with comparators and voltage references. Completely redundant also means redundant power switch transistors. Makes me sleep better. Yeah, I don't trust TI - too amateurish, too many products in too quick cycles, with too little attention to details that matter. Also, tend to need some babysitting.

Yes, I have needed to make a small-scale recall and bodge fix some devices because of a compounded failure of:
1) Me failing SOA calculations in hurry, so that one level of security is wiped out by a shorted MOSFET,
2) TI failing their li-ion chip design, probably in hurry as well, so that the protection which should exist doesn't, and they happily overcharge the cell by actively driving current to it while their internal non-connected signal tells to stop.

This didn't have cell-level measurement - which wouldn't have helped, since the controller was screaming "overvoltage" anyway, cells were in perfect balance; the signal was just ignored. It's the #1 n00b mistake as explained in the Elithion manual, but when it's inside a TI's chip, you can't do anything but add your own external protection. Did I say something about false sense of security? Or babysitting these chips?

--

For your questioning of "nobody would use them" -- it's as I explained before; the chips I'm complaining about are probably "good enough" so they they don't fail all the time, or in much bigger numbers than the devices they are designed in would fail anyway, and when they fail, it's not catastrophic, because the cells handle the abuse because the li-ion R&D business - think about Sony, Panasonic, Samsung, etc. have been really responsible. It's the high tech I'd be proud of, but which gets little credit - only the extremely rare li-ion "explosions" are reported on media!

Now, the funny thing is, when a BMS fails, the whole product fails. When it's (hopefully) examined by the designer for forensics, who is not a li-ion BMS expert, but "just designed in a nice & easy chip", the first thing they do, they measure the cell voltages. A cell is at 0V - so must have been a bad cell! Probably a wrong conclusion. In fact, it was a failed BMS, which killed the cell. The reason gets classified wrong. I'm 99% positive that TI BMS chips fail at 10x or 100x the rate compared to failures originating from the cell. Even 1000x wouldn't surprise me - cell failures are such rare incidences. If cells would randomly short out or start leaking, massive paralleled packs of 18650's, widely used not only by TESLA, would show massive numbers of problems.

The safety is built in the cells. If the cells were really dangerous as people expect, the game would be totally different, and we would have very reliable BMS solutions available out of necessity.

Please understand that TI's just a typical representative example. Others are similar. TI is most widely used and seems to have the largest portfolio. And I think they can handle my critique >:D. And maybe I'm still a bit angry about my 4.63V cells and their partial responsibility about it, and want to vent off?

My point? It isn't to claim these products are total unreliable crap which automatically blow up your battery pack instantly.

It is, people tend to represent these BMS products as completely irreplaceable and extremely robust, extremely well designed safety devices. This is really not the case. Many proper battery systems, typically less than 6s, do not use any cell-level BMS at all.

Yes, I have designed in a TI li-ion I2C MCU AFE once - maybe I was eating the wrong mushrooms while doing that decision? I didn't even go on testing it after getting the prototypes, but desoldered it from the first prototype. It has an I2C interface through which you turn balancing resistors on, then use another message to turn them off. There is no timeout. If this message never comes through - for example, due to I2C bus lockup, which is surprisingly common - you probably have a bricked product. Yeah, implement all the watchdogs. Just to be sure, inject the "typical" I2C reset pattern of clock transitions. How do you test it is effective in actual lockup condition? Any such event during the whole product lifetime could brick it for good. MTBF must be extremely high for no field returns. Really doing this properly on the MCU is a lot of work. So it's not a properly integrated solution - you need considerable work to babysit it. If they just had implemented an utterly trivial timeout counter... But no, every integrated solution seems to have at least one such showstopper-class deficiency.

This feature alone prevents me from using said product. These products are designed for low-quality cheap crap. The performance is similar to what I'd expect from a $0.02 Shenzen special available at LSCS, meant to be used in a toy. What I don't like is the mental image we are given.

But dead products happen! Some percentage of field returns is usually accepted on consumer electronics. In industrial, having a maintenance agreement, and having things break down every now and then (but not too often) can be a good milking cow, depends on how you play the game.

I'm also not proposing any simple answer on "what you should use". Sadly, I'd like to, but I don't know. I don't have a really reliable, good BMS system in my mind. Mine is probably not such; nor is TI's. I revisited the COTS BMS chip offerings this year, spending full working week (60 hours) going through all available BMS products for a 6s battery. None of them made me feel confident.

I expect and hope the safety in battery technology itself to continue increasing so that the BMS would become even less critical. Now, it's already noncritical enough so that the market doesn't need to produce really safe and reliable BMS products.

People want easy answers and quick solutions. Application notes and application field engineers exist for this purpose. This time, I can't give any easy answer - I haven't found it myself. Sorry for that.

TLDR: People just use a TI chip since that's what they are recommended to do. People don't know the failure modes of the batteries. People are happy if a typical product last a typical average lifetime, and returns are within some typical levels.
« Last Edit: October 11, 2018, 07:15:46 pm by Siwastaja »
 

Offline nctnico

  • Super Contributor
  • ***
  • Posts: 27495
  • Country: nl
    • NCT Developments
Sorry for being a bit obnoxious but I'm getting a bit wary when people say 'the whole world has it the wrong way around'. You do make good points. One of the reasons my customer produces battery packs in the NL is because there are almost no decent Li-ion packs available on the market. Most of what comes out of China is too crappy (unreliable) for high-end commercial / industrial use. They do quite a bit of research themselves as well.

When it comes to safety the BMSes I have been dealing with (for high volume production battery packs) ultimately have a fuse which interrupts the circuit when the battery gets shorted and the BMS doesn't cut the power. Other than that there is a big reliance on software and cleverly designed hardware avoiding single points of failure (can't explain further due to confidentiality) where it comes to protection. Other than that battery packs are also required to pass CE and UN38.3 testing. The UN38.3 safety testing is particulary interesting because these involve testing multiple packs for vibration, temperature, charging, discharging. If a pack survives the UN38.3 torture testing you have some degree of a assurance the battery pack won't fall apart or catch fire by itself.
There are small lies, big lies and then there is what is on the screen of your oscilloscope.
 

Offline Siwastaja

  • Super Contributor
  • ***
  • Posts: 8426
  • Country: fi
Sorry for being a bit obnoxious but I'm getting a bit wary when people say 'the whole world has it the wrong way around'.

I agree; OTOH I'm not saying "the whole world has it the wrong way around". It's not that black and white. I'd say, the whole world has it in a non-optimal way, and people design things without thinking. Which I think you'd agree with. It's not limited to li-ion battery management. People do the separate analog and digital ground planes as well as taught by application notes, without giving it a thought. Now, there is one difference: no one is saying that your things catches fire and explodes by using the wrong kind of ground plane. This kind of simplified argumentation is true when discussing about li-ion management.

Quote
there are almost no decent Li-ion packs available on the market. Most of what comes out of China is too crappy (unreliable) for high-end commercial / industrial use.

We noticed exactly the same. Spawning our own cell testing system to verify/prove/disprove assumptions and, foremost, evaluate real-life cycle lifes in different conditions. Trying to find "optimum way", I built this in 2014-2015: . It's still in use, I now use it for building battery packs for mobile robots in a related startup I now design for... Not using nickel strip but direct copper interfaces is both cost and performance optimization.

Quote
When it comes to safety the BMSes I have been dealing with (for high volume production battery packs) ultimately have a fuse which interrupts the circuit when the battery gets shorted and the BMS doesn't cut the power.

Yes. The bog standard fuse. Rated correctly, it's the most important protection. (Note that some BMS's have some peculiar limitations when adding fuses or contactors mid-pack. Always be sure to understand these limitations.)
 


Share me

Digg  Facebook  SlashDot  Delicious  Technorati  Twitter  Google  Yahoo
Smf