Does my car pollute?

Does my car pollute?

Anti-pollution measures, speed limits, parking restrictions, even the grey sky colour, and very, very alarming data. These are the consequences of the circulation of our cars in big cities. According to the European Environment Agency (EEA), more than 13% of the polluting particles in the 28 countries of European Union are produced by transport, which supposes almost 4.000 deaths per year. Only in cities, data ensure that traffic produces the 60% of emissions to the atmosphere. How long can we continue allowing this situation?

However, not all cars are so guilty of these emissions. Only 10% of the vehicles that circulate in our streets contribute 50% of the emissions, according to experts. They are what we call “high emitters” (HE). But, which are these cars? Diesel engines? The oldest ones? The worst maintained by theirs owners? Not necessarily. A high percentage of owners of highly polluting vehicles are not aware of it. Many of them have successfully passed the vehicle inspection and even 50% of these high emitters have less than two years.

How can we find out if our car is a “high emitter”?

LIFE GySTRA project, coordinated by CARTIF, purposes identifying this kind of highly polluting vehicles and monitor continuously the evolution of empiric emissions levels to quantify the savings of emission volumes. This process will be possible thanks to a new technological development, the RSD +. For the moment, the intention is to carry out tests and collect data in order to launch a new sustainable mobility policy.

The demonstration study will be carried out in Madrid (Spain) and Sofia (Bulgaria), where the intention is to control the vehicles that circulate in both cities thanks to three RSD + devices, adapted to the requirements of the EU in terms of NO2 emission control.

The public model in Madrid (Spain) is going to monitor 700,000 vehicles per year, with two RSD+ devices. The owners of the vehicles identified as HE will be notified to proceed with car reparation. With the repair of this kind of cars, it is expected to achieve emission savings of 14.8% (CO) and 22.7% (NOx, NO and NO2) of the total volume of emissions. If only the half of the total HE is repaired it would be possible to reduce CO2 emissions up to 16Mt per year.

On the other hand, the fleet model of Sofia (Bulgaria) is going to control a fleet integrated by 150 buses continuously measured. A recent study on buses concluded that identifying 6.6% of HE and repairing them their emissions were reduced by up to 84%. This monitoring program will allow higher emission savings, and fuel savings are expected to be 3-5% for the HE.

The repair of these vehicles does not only mean environmental advantages, but it will mean economic savings and the improvement of vehicle conditions.

If the project team achieves these objectives, it will greatly reduce pollution in our cities, even reaching to avoid episodes of high pollution and the restrictions, which mean headache for citizens and administrations.

The project is designing too an emission reduction policy that includes information campaigns aimed at population, some more general and others specific to the owners of the most polluting vehicles.

The project consortium is integrated by five partners, three of them technological and two from the administration. Firstly, CARTIF coordinates the proposal; OPUS RSE is the company that will develop RSD+ technology for remote contamination monitoring; and CIEMAT, the research centre that will calibrate the equipment and perform the characterization and evaluation of emissions. On the other hand, the Spanish Traffic General Direction and the City Council of Sofia (Bulgaria) will lend their support for the demonstration study in the cities of Madrid and Sofia, respectively.

Strategic urban planning

Strategic urban planning

According to the United Nations, in 2014 more than half of the world’s population was living in urban areas and two third of the world’s population will be living in an urban area by 2050 being Europe the most urbanized continent (URBACT, 2015). The forecast for 2050 in this case is that the percentage will increase up to 75% (Eurostat, 2016). Besides, urban areas are engines of regional and national growth as they generate 53% of gross national product (GNP) in low-income countries, 73% in middle-income countries, and 85% in high-income countries (World Bank, 1999).

Although the concentration in cities usually supposes an increase of density and less consumption of resources, cities use two-thirds of the world’s energy and generate three-fourths of the world’s CO2 emissions (Smart Cities Council, 2013). In addition, urban areas have important drawbacks, those being waste production, carbon emissions, pollution, lack of preservation of heritage and environment, traffic congestion, etc.

The traditional urbanism has not been able to give response to the current situation and problems that have arisen in recent years in cities, due to its complexity, diversity and uncertainty. Therefore, new planning instruments are needed, such as Strategic Planning, as an attempt to address the complexity and socio-economic diversity of our cities from a multidisciplinary perspective.

According to the definition of professor Fernández Güell, Strategic Urban Planning is “… a systematic, creative and participatory process that stablishes the bases of an integrated long-term vision, which defines the future development model, which formulates strategies and actions to achieve this model, which enables a continuous decision-making system that involves local agents throughout the entire process”.

To sum up, it is a deep study of the cities, in order to understand their current state, where are we?, understanding the past so as to help us to understand the present: where do we come from?, and, finally, to define a city model or future vision in accordance with political and citizen aspirations: where do we want to go?

Why is a city strategy necessary nowadays?

It is necessary as a way to achieve the sustainable urban development, understood from the three points of view: environmental, socioeconomic and institutional, as a global and systemic approach. This is considered as the starting point for the definition of the baseline situation of the city as well as the future vision. The current city model, as well as the way of life of its citizens need to be reconsidered, and cities need to find a way to regenerate themselves, with the aim of ensuring sustainability in the medium-long term, as well as being able to meet the challenges mandated by the European Commission by 2020 or by 2030 (40% reduction of greenhouse gas emissions (in relation to 1990 levels), 27% share of renewable energy, and 27% energy efficiency improvement).

Therefore, Strategic Urban Planning has become the best instrument to tackle the challenges that cities are currently facing.

What does CARTIF do in this regard?

From CARTIF within smart city projects, we collaborate with European cities that want to reformulate their city model into a model leveraging the convergence of energy, mobility and ICT to transform European cities into Smart Cities, through the development of an integrated strategy.

As for example within MAtchUP project, we are working in the development of an integrated strategy with the cities of Valencia, Dresden and Antalya, or in MySMARTLife project with Hamburg, Helsinki and Nantes.

Aiming a wider approach, we are currently working with municipalities such as Laguna de Duero (Valladolid, Spain) for the definition of its Strategic Plan 2017-2022, which will serve as guidance for municipal interventions and policies in the coming years.

Geographic Information Systems and their application in the energy field

Geographic Information Systems and their application in the energy field

Even though the term Geographic Information System (GIS) is well-known, it is possible that many of you don’t know what applications it might have or its relevance in the energy field. Put in short, GIS (or SIG, in Spanish) are all software in charge of the treatment of data containing some geometric characteristic and that can be reflected on a map in their precise position. These data can be 2D or 2,5D* (described with points, lines and polygons), 3D, or cloud points (LIDAR data). Moreover, these geographic data are normally associated to attribute tables, where information on them is introduced. For example, we can have a map with the provinces in Spain and in the attribute table have assigned to each polygon representing a province their demographic data, economic data, etc.

One of the most remarkable aspects of these systems is not only being able to visualize elements in their precise geographic location, but also that these layers of information can be overlapped allowing to visualize at the same time geographical elements displaying different realities. This is quite straightforward and we are very used to seeing it in phone apps, for example GPS apps, where we can observe a base map (a city map or a satellite image) and several layers that are placed on top of it, such as the name of the streets, stores, etc.

A part from being able to use these systems in order to guide ourselves in cities (which is no small thing) the potential of these systems lies in being able to perform spatial analysis, which would be impossible with other means. This way, we could have answers to questions like the following:

  • What would be the floodable areas by this river?
  • If an incident occurs in this area, which are the closest hospitals? What would be the best route for ambulances with respect to distance? And with respect to time?
  • Where should the stops of this bus line be placed in order for them to be spaced at a maximum of 600 meters? Which areas in the city would benefit from it considering a radius from the bus stop of 10 minutes walking?
  • How have forest areas been modified in a concrete zone? Is there risk of desertification?

These only represent a small sample of the reach of GIS, which proves extremely useful to carry out planning activities in a wide range of fields (risks and accidents, traffic management, transport networks, environmental impact assessment, agriculture, natural risk assessment…). But focusing on the energy field, GIS have also a great potential for the support in the development of energy plans, compliance with energy directives and result monitoring. For example, we could get to know which areas are in need to perform an energy retrofit. To this respect it is worth mentioning as an example the map developed by the University of Columbia on the estimated consumption in New York City.

Additionally, several different scenarios can be evaluated where the effectiveness of the different actions is measures or if a determined area can be supplied by other type of energy source (renewable, for example). Calculating these indicators it can be checked if the objectives imposed in a determined directive are complied with or not.

In CARTIF, and in particular in the Energy Division, GIS are exploited and their applications to support to the compliance with the European Directives in the energy field, more specifically to the Directive package “Clean Energy for All Europeans”. Moreover, special attention is paid to the study of the data structure and the standards that should be followed to assure its interoperability. In this sense, it is worth highlighting the open standards proposed by the Open Geospatial Consortium (OGC), and also the INSPIRE Directive, which defines the infrastructure for the spatial information in Europe and which will be applicable in 2020.

This latter aims at harmonising and offering geospatial information in Europe in a range of 34 themes. Even though none of them is strictly related to energy (these aspects can be assigned to build elements, such as buildings (BU)), the study of the most relevant energy attributes is crucial in this moment prior to the implementation of the INSPIRE Directive, as it has been manifested by the European Commission when defining a project that studies the potential of the Directive in the energy field: the “Energy Pilot”. CARTIF, aiming for innovation and the alignment with the EU collaborates in this project interacting with one of the reference centres of the Commission: the Joint Research Centre in Ispra.

*Note for the curious: for example a cube can be considered 2,5D when it is defined instead of with eight vertexes with x, y and z values, it is defined only with the four above, since those contain the “z” value in contrast to the four lower vertexes, where this value would be 0.

Are we ready for electric self-consumption?

Are we ready for electric self-consumption?

That the self-consumption is a real option and that is going to grow little by little is evident. But are we really prepared? The next few lines describe some steps to establish it in our homes.

In the subject of self-consumption there is a concept that we must never forget: energy efficiency. This efficiency must be understood from both the generation and consumption sides.

Let us first analyze efficiency from the point of consumer. It is evident that if my household consumes less electricity, the cost of my self-consumption facility will be lower. Are we taking any measure of efficiency for this to occur? A first step that can be taken is to reduce the consumption of lighting in the home. The change of halogen and low consumption bulbs by others of LED technology will allow us to reduce enough the electricity consumption in lighting. Another step that we can take is to replace our old appliances with others of class A +++ that have a lower level of consumption.

Efficient measures that are not always within reach of most budgets are to improve the isolation of our home. The insulation of the envelope of the building is fundamental. The use of insulation in facades, ceilings and floors and a suitable choice of windows can reduce the consumption of our building.

Other measures simply go through the change of habits in consumption that we must learn if we want to implement self-consumption in the home. The simple gesture of turning off light bulbs or electrical appliances that are not used, to avoid to leave electronic devices in stand.by (phantom energy) and to put the appliances in operation in the hours of the day when more energy is generated will allow an efficient management of our system. This can be done by implementing energy management software (EMS) in our home but it is an added cost.

If we are thinking about buying an electric car maybe this is the time to choose it with V2G (Vehicle to grid) technology with its Vehicle to Home (V2H) and Vehicle to Building (V2B) variants. This technology allows energy stored in an electric car to be injected into the electrical grid or to a dwelling or building using the car battery as an electrical storage system. In this way a better integration of renewable energies can be achieved in the electrical system.

Perhaps these measures will allow a home to consume only 1500 kw/h a year compared to the current 3000 kw/h of average consumption in Spain. This would reduce the cost of our self-consumption facility which would mean that many households will consider doing this installation in their homes.

From the point of view of the generation side, progress is being made by leaps and bounds. The efficiency of existing photovoltaic panels that use new materials with a longer useful life is very far from those manufactured 10 years ago and the price per w is lower, reaching values of 0.8 € per watt installed. Equally, the technology of the batteries makes them more efficient and with a greater durability supporting greater cycles of recharge and at a lower price.

And is the electricity grid ready for self-consumption? According to the operator of the Spanish electricity system the network is prepared for hundreds of thousands of self-consumers to enter the network.

What are the electric companies doing? Power companies are realizing that self-consumption will sooner or later arrive to settle permanently in each of our homes and that is the time has come to move ahead. Some companies start to market self-consumption kits, control systems or maintenance contracts that ensure a proper functioning of the system.

What is necessary so that everything starts to work? As easy as getting to an agreement point where distributors (power companies) begin to see prosumers (current users and future producers) as potential allies and non-competitors.

On the one hand, the electricity companies claim that the use of the distribution network should be paid not only by the current consumer but by the future producer and as we know these costs represent more than 50% of the current electricity tariff. But it is also true that companies are going to save the generation costs that are difficult to know at present.

But what would happen if a large number of users decide to become only their own power generators and definitely disconnect from the network? And it is then when the government enters and depending on the laws that apply and according to on whether they are advantageous for the consumer, the companies or both when the consumer decides.

How to improve your processing plant without large investments?

How to improve your processing plant without large investments?

Any processing plant – continuous, batch, hybrid – can improve its economic, safety and environmental indicators in two ways: improving its processing equipment or improving the control of those equipments.

The improvement of processing equipment is usually a task that requires large investments, since it almost always involves the acquisition of new processing equipment or in the best case requires expensive remodeling.

On the contrary, these performance indicators can be substantially improved through control, without, in the majority of cases, any investment in new technical means of instrumentation and control. This is because in practically all cases there is a wide margin to improve the performance indicators of a processing plant through its regulation.

The origin of this margin of improvement is multiple. The most common causes are: the control system is not well designed or tuned; due to ignorance or haste, all the benefits of the available control system are not used; the automaton programmer or the process engineer are not control experts; the dynamics of the processes under automatic regulation are not known with the required depth; the design of the plant has not been made under the integrated design approach.

Are also diverse and numerous actions that can be implemented to improve the performance of control loops without any investment and that we will review in next posts, such as: improve the tuning of the regulator, redesign the controller, implement anticipatory compensation of disturbances, enhance the tuning of cascade regulation loops, redesigning or re tuning the level controllers of the buffer tanks if necessary, using a control algorithm advanced available or supportable by the controller instrument, reducing couplings between loops, making a better assembly of the measuring probe, etc.

In its vast majority, the processing plants are automated under control structures (basic control, cascade control, split range control, selective control, coupled loops, etc.) based on the universal PID controller, in all its particularizations (P, PI, PD, PID, PI_D, etc.).

Despite its longevity and the development of multiple advanced control techniques, PID control maintains an overwhelming presence in the process industry.

Its extensive use in the industry is such that all the surveys known by the author conclude unanimously, in which more than 95% of the existing control loops are of the PID type. However, also many surveys conclude that a high percentage of the loops with PID control in the world, are operated in manual mode, while another similar percentage operates defectively. For example, as shown in the following figure, in [1] it is reported that only 16% of the PID regulation loops are optimally tuned and their performance is therefore excellent.

There is no doubt that in most cases, the incorrect or poor tuning of the controller can be the cause of the poor performance of the control loop or its irregularities in the operation.

However, it should not be forgotten that automatic regulation systems are holistic systems, and as such they must be analyzed as a whole and not only through the parts that compose them. That is why it is necessary to review the other components of the loop before deciding what action should be exercised on said loop.

Hence, the procedure of action in all cases, must begin with a field review of all the components of the loop (controller, process, actuator, measurer and communication channels), as well as an analysis of the possibility of coupling with other process loops.

The result of this first phase will determine what concrete action corresponds to perform to solve the poor performance of the automatic regulation loop.

CARTIF offers this service to optimize the performance of the regulation systems of processing plants. The optimization reduces the oscillations and the variability of the production plant, making the regulation system more accurate, faster, more stable and safer, and in this way improving its efficiency, safety, environmental impact and profitability.

In next post, the execution procedure will be described for each of the possible actions, starting with the simplest one, the re-tuning of the controller.

New applications of Deep Learning

New applications of Deep Learning

A little more than a year ago, in another post of this blog, our colleague Sergio Saludes already commented what is deep learning and detailed several of its applications (such as the victory of a machine based on these networks over the world champion of Go, considered the most complex game in the world).

Well, in these 16 months (a whole world in this topic) there has been a great progress in terms of the number of applications and the quality of the results obtained.

Considering, for example, the field of medicine, it has to be said that diagnostic tools based on deep learning are increasingly used, achieving in some cases higher success rates than human specialists. In specific specialties such as radiology, these tools are proving to be a major revolution and in related industries such as pharmaceuticals have also been successfully applied.

In sectors as varied as industrial safety, they have recently been used to detect cracks in nuclear reactors, and have also begun to be used in the world of finance, energy consumption prediction and in other fields such as meteorology and the study of sea waves.

Autonomous vehicle driving projects, so in vogue these days, mainly use tools based on deep learning to calculate many of the decisions to be made in each case. Regarding this issue, there is some concern about how these systems will decide what actions to take, especially when human lives are at stake and there is already a MIT webpage where the general public can collaborate in creating an “ethics” of the autonomous car. Actually, these devices can only decide what has previously been programmed (or trained) and there is certainly a long way to go before the machines can decide for themselves (in the conventional sense of “decide”, although this would lead to a much more complex debate on other issues such as singularity).

Regarding the Go program discussed above (which beat the world champion by 4 to 1), a new version (Alpha Go Zero) has been developed that has beaten by 100 to 0 to that previous version simply knowing the rules of the game and training against itself.

In other areas such as language translation, speech comprehension and voice synthesis have also advanced very noticeably and the use of personal assistants on the mobile phone is beginning to become widespread (if we overcome the natural rejection or embarrassment of “talking” with a machine).

CARTIF is also working on deep learning systems for some time now and different types of solutions have been developed, such as the classification of architectural heritage images within the European INCEPTION project.

All these computer developments are associated with a high computational cost, especially in relation to the necessary training of the neural networks used. In this respect, progress is being made on the two fronts involved: much faster and more powerful hardware and more evolved and optimized algorithms.

It seems that deep learning is the holy grail of artificial intelligence in view of the advances made in this field. This may not be the case and we are simply looking at one more new tool, but theres is no doubt that is an extremely powerful and versatile tool that will give rise to new and promising developments in many applications related to artificial intelligence.

And of course there are many voices that warn of the potential dangers of this type of intelligent systems. The truth is that it never hurts to prevent the potential risks of any technology, although, as Alan Winfield says, it’s not just artificial intelligence that should be feared, but artificial stupidity. Since, as always happens in these cases, the danger of any technology is in the misuse that can be given and not in the technology itself. Faced with this challenge, what we must do is promote mechanisms that regulate any unethical use of these new technologies.

We are really only facing the beginning of another golden era of artificial intelligence, as there have been several before, although this time it does seem to be the definitive one. We don’t know where this stage will take us, but trusting that we will be able to take advantage of the possibilities offered to us, we must be optimistic.