Category: Blog

AQ: Determine coefficient of grounding

Determination of required grounding impedance is based on determination of coefficient of grounding which represents ratio of maximum phase voltage at phases which aren’t exposed by fault and line voltage of power network:

kuz=(1/(sqrt(3)))*max{|e(-j*2*π/3)+(1-z)/(2+z)|; |e(+j*2*π/3)+(1-z)/(2+z)|}
z=Z0e/Zde

where are:

kuz-coefficient of grounding,
z-ratio of equivalent zero sequence impedance viewed from angle of place of fault and equivalent direct sequence impedance viewed from angle of place of fault,
Z0e-equivalent zero sequence impedance viewed from angle of place of fault,
Zde-equivalent direct sequence impedance viewed from angle of place of fault.

So, after this explanation, you can get next conclusions:
if kuz=1 then power network is ungrounded because Z0e→∞, which is a consequence of existing more (auto) transformers with ungrounded neutral point than (auto) transformers with grounded neutral point (when kuz=1 then there aren’t (auto) transformers with grounded neutral point),
if kuz≤0,8 then power network is grounded because Z0e=Zde, which is a consequence of of existing more (auto) transformers with grounded neutral point than (auto) transformers with ungrounded neutral point.

Fault current in grounded power networks is higher than fault current in ungrounded power networks. By other side, in case of ungrounded power networks we have overvoltages at phases which aren’t exposed by fault, so insulation of this conductors could be seriously damaged or in best case it could become older in shorter time than it is provided by design what is the main reason for grounding of power networks.
Coefficient of grounding is very important in aspect of selecting of insulation of lighting arresters and breaking power of breakers, because of two next reasons:
1. in grounded power networks insulation level is lower than insulation level in ungrounded power networks,
2. in grounded power networks value of short circuit current is higher than value of short circuit current in ungrounded power networks.

AQ: How to design an Panel required for PLC / MCC / Drive

1. The regular industrial standard size panel available with most of the panel fabricator’s.
2. Type of protection (used to say as IP).
3. Spacing depends upon the Power handled by the conductors inside Panel and the ventilation system.
4. Cable Entry / Bus bar entry may depend on the application and site condition. it may be at rear/bottom or at the top.
5. When comes to Drive, if the site condition is too hot then an industrial ac is required usually attached at the side of the panel.
6. Drive to drive required spacing (Check the manual of the drive}, since the power switching activity take place inside the drive.
7. Provide required space for the transformers and AC-choke since they create magnetic flux in ac circuits.
8. Don’t mix the Control cable, Power cable, Signal cable and Communication cable together in the cable tray… Otherwise you will be wired…
9. Keep the control on mcb/mccb/mpcb in handy location. So that its easier for operator to control it frequently and not disturbing other circuits..
10. Plc will be acquiring the top position in the panel since there is nothing to do with it once installed. Just we will be monitoring the status.
11. Don’t place the Plc nearby to the incoming or outgoing heavy power terminals..
12. Mcc panel are easy thing to do, but do the exact calculation for the ACB selection in the incomer side. Since each feeder will be designed with tolerance level.

There being a lot more than 12 guidelines to follow. What
about back-up power for the PLC? What about internal heat flow considerations
(not just does it need an AC or not)? How much space between terminal blocks
and wireway? What about separate AC and instrument grounds? What about wireway
fill? What about wire labels? What about TSP shields? What about surge
protection? In my experience, there are plenty of people that can design a
panel but if they haven’t gone to the field with it then they haven’t been able
to learn from their design mistakes.

The best thing you can do is start your design but you
really need to be guided by an experienced designer.

AQ: Right Half Plane Pole

Very few know about the Right Half Plane Pole (not a RHP-Zero) at high duty cycle in a DCM buck with current mode control. Maybe because it is not really a problem.
It is said that this instability starts above 2/3 duty cycle – I think that must be with a resistive load. If loaded with a pure current source, it starts above 50% duty cycle.

Here is a little down-to-earth explanation:
If you run a buck converter at high duty cycle but DCM, it probably works fine and is completely stable. Then imagine you suddenly open the feedback loop, leaving the peak current constant and unchanged. The duty cycle will then rush either back to 50% or to 100% if possible. You now have a system with a negative output resistance – if Voltage goes up, the output current will increase.

You can see it by drawing some triangles on a piece of paper: A steady state DCM current triangle with an up-slope longer than the down-slope and a fixed peak value. Now, if you imagine that the output voltage rises, you can draw a new triangle with the same peak current. The up-slope will be longer, the down-slope will be shorter but the sum of times will be longer than in the steady state case. The new triangle therefore has a larger area than the steady state triangle, which means a higher average output current. So higher output voltage generates higher output current if peak current is constant. Loaded with a current source, it is clear that this is an unstable system, like a flipflop, and it starts becoming unstable above 50% duty cycle.

However, when you close the feedback loop, the system is (conditionally) stable and the loop gain is normally so high at the RHP Pole frequency that it requires a huge gain reduction to make it unstable.

It’s like when you drive on your bike. A bike has two wheels and therefore can tilt to either side – it is a system with a low frequency RHPP like a flipflop. If you stand still, it will certainly tilt to the left or to the right because you have no way to adjust your balance back. But if you drive, you have a system with feedback where you can immediately correct imbalance by turning the handlebars. As we know, this system is stable unless you have drunk a lot of beers.

AQ: What causes VFD driven motor bearing current?

There are several things involved, all with varying degrees of impact.

Large machines are – generally speaking – made of pieces (segments) because the circle for the stator and/or rotor core is too large to manufacture from a single sheet. This leads to some breaks in the magnetic flux path symmetry, both in the radial (right angles to the shaft) and axial (parallel to the shaft) directions.

For the most part, the windings of large machines are formed and installed by hand. This too can lead to symmetry issues, as the current paths are not identical which in turn will create some differences in the magnetic field flux.

Output waveforms from power electronics are only approximations of true sinusoids. The presence of additional harmonics distorts the sinusoidal nature and results in changes that are not symmetric in the magnetic field strength … which in turn means a non-symmetric flux distribution.

Two other items contribute to potentially damaging bearing currents as well. One of these is the Common Mode Voltage which is present (to some degree) in all drives. Essentially this is a signal that is present at both the drive input and output … I tend to think of it as an offset. It’s not something that traditional grounding addresses, and can create an elevated potential in the shaft which then discharges through the bearing path.

A second item is not related to the presence (or absence) of drives at all; it is related to the mechanical arrangement of the process drive train. For example, a shaft that has a sliding seal (like the felt curtain on a dryer section), or one that turns a blade against a gas or liquid (like a compressor) can generate a static charge at the point of contact. If there is no means of isolating this charge to the portion of the shaft where the sliding is occurring, it can pass through to the motor shaft and thence through the motor bearings.

Lastly – the frequency of the variable frequency drive harmonics in the output waveform is significantly higher than line frequency. This requires specific accommodations for grounding as traditional methods are insufficient due to the attenuation caused by the relatively high resistance ground path.

AQ: Avoid voltage drop influence

My cable size and transformer size should give me maximum 3% on the worst 6% to 10%. If it is the single only equipment on the system then maybe you can tolerate 15%. If not, dip factor may affect sensitive equipment and lighting.

This is very annoying for office staff each time a machine starts lights are dimming. It does not matter what standard you quote I cannot accept 10%- 15% make precise calculation and add a 10% tolerance to avoid.

In most cases, this problem comes from cable under sizing so we have to settle with a Standard giving 15% Max.

Just recently I had to order a transformer and cable change for a project which was grossly undersized.
I have had to redesign the electrical portion of a conveyor and crushing system to bring the system design into compliance with applicable safety codes. The site was outdoor at a mine in Arizona where ambient temperatures reach 120F. The electrical calculation and design software did not include any derating of conductor sizes for cable spacing and density within cable trays, number of conductors per raceway, ambient temperature versus cable temperature rating, etc. Few of the cables had been increased in size to compensate for voltage drop between the power source and the respective motor or transformer loads.

Feeder cables to remote power distribution centers were too small, as voltage drop had not been incorporated in the initial design. The voltage drop should not be greater than 3%, as there will be other factors of alternating loads, system voltage, etc. that may result in an overall drop of 5%.

The electrical system had to be re-designed with larger cables, transformer, MCCS, etc, as none of the design software factors in the required deratings specified in the National Electric Code NFPA70 nor the Canadian Electric Code, which references the NEC.

AQ: SCADA & HMI

SCADA will have a set of KPI’s that are used by the PLCs/PACs/RTUs as standards to compare to the readings coming from the intelligent devices they are connected to such as flowmeters, sensors, pressure guages, etc.

HMI is a graphical representation of your process system that is provided both the KPI data and receives the readings from the various devices through the PLC/PAC/RTUs. For example you may be using a PLC that has 24 i/o blocks that are connected to various intelligent devices that covers part of your water treatment plant. The HMI software provides the operator with a graphical view of the treatment plant that you customize so that your virtual devices and actual devices are synchronized with the correct i/o blocks in your PLC. So, when an alarm is triggered, instead of the operator receiving a message that the 15th i/o block on PLC 7 failed, you could see that the pressure guage in a boiler reached maximum safety level, triggering a shutdown and awaiting operator approval for restart.

Here is some more info I got from my colleague who is the expert in the HMI market, this is a summary from the scope of his last market study which is about a year old.

HMI software’s complexity ranges from a simple PLC/PAC operating interface but as plant systems have evolved, HMI functionality and importance has as well. HMI is an integral component of a Collaborative Production Management (CPM) system; simply you can define that as the integration of Enterprise, Operations, and Automation software into a single system. Collaborative Production Systems (CPS) require a common HMI software solution that can visualize the data and information required at this converged point of operations and production management. HMI software is the bridge between your Automation Systems and Operations Management systems.

An HMI software package typically performs functions such as process visualization and animation, data acquisition and management, process monitoring and alarming, management reporting, and database serving to other enterprise applications. In many cases, HMI software package can also perform control functions such as basic regulatory control, batch control, supervisory control, and statistical process control.

“Ergonometrics,” where increased ergonomics help increase KPI and metric results, requires deploying the latest HMI software packages. These offer the best resolution to support 3D solutions and visualization based on technologies such as Microsoft Silverlight. Integrating real-time live video into HMI software tools provide another excellent opportunity to maximize operator effectiveness. Live video provides a “fourth dimension” for intelligent visualization and control solutions. Finally, the need for open and secure access to data across the entire enterprise drives the creation of a single environment where these applications can coexist and share information. This environment requires the latest HMI software capable of providing visualization and intelligence solutions for automation, energy management, and production management systems.

AQ: AM & FM radio

For AM & FM radio & some data communications adding the QP filter make sense.
Now that broadband, wifi, data communications of all sizes & flavours exist – any peak noise is very likely to cause interuptions & loss of integrity of data – all systems are being ‘cost reduced’ ensuring that they will be more susceptible to noise.
I can understand the reasons for the tightening of the regulations.
BUT, it links in to the other big topic of the moment – the non-linearity of managers.
William is obviously his own manager – I bet if his customer was to ask him to spend an indefinite amount of time fixing all the root causes to meet the spec perfectly without any additional cost it would be a different matter.

Unfortunately for most of us the realities of supervisors wanting projects closed & engineering costs minimized we have to be careful in the choice of phrasing.
Any suggestion that one prototype is ‘passing’ suddenly can be translated to job finished, & even in our case where the lab manager mostly understands, his boss rarely does & the accountant above him – not at all.

It gets worse than that – at the beginning of a project (RFQ) – the question is “how long will EMC take to fix?” with the expectation if a deterministic answer; the usual response of a snort of derision & how long is a piece of string generally translates to 2 weeks & once set in stone becomes a millstone (sorry mile-stone).

We already have a number of designs that while not intentionally using dithering, do use boundary mode PFC circuits which automatically force the switch frequency to vary over the mains cycle. These may become problematic at some future variation of the wording of the EMC specs.

While I have a great deal of sympathy for the design it right first time approach, the bottom line for any company is – it meets the requirement (today) – sell it!!

AQ: Electronic industry standards

You know standards for the electronic industry have been around for decades, so each of the interfaces we have discussed does have a standard. Those standards may be revised but will still be used by all segments of our respective engineering disciplines.

Note for example back in the early 1990s many big companies HP, Boeing, Honeywell … formed a standards board and developed the Software standards( basic recommendations) for software practices for programming of flight systems. It was not the government it was the industry that took on the effort. The recommendations are still used. So an effort is first needed by a meeting of the minds in the industry.

Now we have plenty of standards on the books for the industry, RS-422, RS-232, 802.1 … and the list goes on and on. The point is most of the companies are conforming to standards that may have been the preferred method when that product was developed.

In the discussion I have not seen what the top preferred interfaces are. I know in many of the developments I have been involved in we ended up using protocol converters, Rs-232 to 802.3, 422 to 485 … that’s the way it’s been in control systems, monitoring systems, Launch systems and factory automation. And in a few projects no technology existed for the interface layer, had to build from scratch. Note the evolution of ARPA net to Ethernet to the many variations that are available today.

So for the short hall if I wanted to be more comparative I would use multiple interfaces on my hardware say usb, wireless, and 422. Note for new developments. With the advancement in PSOCS and other forms of program logic interface solutions are available to the engineer.

Start the interface standards with the system engineers and a little research on the characteristic of the many automation components and select the ones that comply with the goals and the ones that don’t will eventually become obsolete. If anything, work on some system standards. If the customer is defining the system loan him a systems engineer, and make the case for the devises your system or box can support, if you find your product falls short build a new version. Team with other automation companies on projects and learn from each other. It’s easy to find issues as to why you can’t succeed because of product differences, so break down the issues into manageable objectives and solve one issue at a time. As they say divide and concur.

AQ: Spread spectrum of power supply

Having lead design efforts for very sensitive instrumentation with high frequency A/D converters with greater than 20-bits of resolution my viewpoint is mainly concerned about the noise in the regulated supply output. In these designs fairly typical 50-mV peak-to-peak noise is totally unacceptable and some customers cannot stand 1-uVrms noise at certain frequencies. While spread spectrum may help the power supply designer it may also raise havoc with the user of the regulated output. The amplitude of the switching spikes (input or output) as some have said, are not reduced by dithering the switching frequency. Sometimes locking the switching time, where in time, it does not interfere with the circuits using the output can help. Some may also think this is cheating but as was said it is very difficult getting rid of most 10megHz noise. This extremely difficulty applies for many of the harmonics above 100kHz. (For beginners who think that being 20 to 100 times higher than the LC filter will reduce the switching noise by 40 to 200 are sadly wrong as once you pass 100kHz many capacitors and inductors have parasitics making it very hard to get high attenuation in one LC stage and often there is not room for more. More inductors often introduce more losses as well.) We should be reducing all the noise we can and then use other techniques as necessary. With spread spectrum becoming more popular we may soon see regulation on its total noise output as well.

One form of troublesome noise is common mode noise coming out of the power inputs to the power supply. If this is present on the power input to the power supply it is very likely it is also present in the “regulated” output power if floating. Here careful design of the switching power magnetics and care in the layout can help minimize this noise enough, that filters may be able to keep the residual within acceptable limits. Ray discusses some of this in his class but many non-linear managers frequently do not think it is reasonable or necessary for the power supply design engineer to be involved in layout or location of copper traces. Why not, the companies that sell the multi-$100K+ software told their bosses the software automatically optimizes and routs the traces.

Spread spectrum is a tool that may be useful to some but not to all. I hope the sales pitch for those control chips do not lull unsuspecting new designers into complacency about their filter requirements.

AQ: Home automation concept

The concept of home automation on a global scale is a good concept. How to implement such a technology on a global scale is an interesting problem, or I should say issues to be resolved. Before global approval can be accomplished the product of home automation may need a strategy that starts with a look at companies that have succeeded in getting global approval of their products.

If we look at what companies that have the most products distributed around the world we see that Intel is one of these companies. What’s interesting is that this company has used automation in their Fabs for decades. This automation has allowed them to produce their products faster and cheaper than the rest of the industry. The company continues to invest in automation and the ability to evolve with technology and management. We have many companies that compete on the world stage; I don’t think many of these companies distribute as much product. So to compete at a level to make home automation accepted and to accomplish global acceptance the industry and the factories have to evolve to compete. That mission by the automation can be accomplished by adapting a strategy that updates their automation in their factories, stop using products that were used and developed in the 1970s (another way of saying COTS) and progress to current and new systems. A ten years old Factory may be considered obsolete if the equipment inside is as old as the factory.

Now for cost, when I thank of PLC or commercial controllers I see a COTS product that may be using obsolete parts that are not in production any more or old boards. So I see higher cost for manufacturing, a reduction in reliability. Now many procurement people evaluate risk in such a way that may rate older boards lower in risk for the short term, not a good evaluation for the long term. The cost is a function of how much product can be produced at the lowest cost and how efficient and competitive the company that produces the product. So time is money. The responsibility for cost is the company and the ability to produce a competitive product, not the government.

Now into control systems and safety, if the automation system is used in the house safety has to be a major consideration. I know at Intel Fabs if you violate any safety rule you won’t be working at that company long. To address safety the product must conform to the appropriate standards. Safety should be a selling point for home automation. Automation engineers should get and remember safety is one of the main considerations for an engineer. If someone gets hurt or killed because of a safety issue the first person looked at is the engineer.

Now 30% energy saving in my book is not enough, 35 to 40 percent should be a goal. Now solar cells have improved but the most efficient in the south west US. The Sterling engines are 1960 designs and use rare gases such as helium which may not be a renewable resource, Wind generators need space and are electromechanical so reliability and maintenance needs improving.

Now on to the interface standards, most modern factories that produce processors use the Generic equipment Manufacture standard, good deal works. As far as what and when to uses a standard interface, on BOX produced by one company may use RE-422 where another company may use RS 485 so the system engineer should resolve these issues before detailed design starts. Check with IEEE. Or you may be able to find the spec at every spec.com this is a good place to look for some of the specs needed.

So I conclude, many issues exist, and when broken down home automation is viable and needs a concerted effort and commitment from at least the companies and management that produce products for automation and a different model for manufacturing and growing the home systems.
Home automation with a focus on energy savings as a goal is a good thing. We have a lot of work to ma