>> Top Banking analytics companies
Smart cities are considered as the future of urban living, using digital technologies, data, and design thinking. As the urban population grows in diverse conditions, many cities turn to technology and smart city solutions to build a more livable environment and thereby improving the delivery of public services.
Smart cities create a new model by improving existing transportation systems, that use the data applications and technology to improve the quality of life for citizens and visitors. Technology serves citizens and can enhance many aspects of public life, including quality of life, mobility, security, education, economy, and the environment. Here are some key points that enable the smart government of smart enterprises movement globally.
In this race for applying AI in unique and attractive ways, an important piece of the AI learning curve is missing—the flashy AI technology used at the front end cannot be fully appreciated without a complex infrastructure that comprises end-to-end process automation. Read More
Future of Dubai in 2021 Ever wondered how dubai got such improvement in such a short span of time. It was all possible due to its goal of being number 1 in all sector.
Share your opinion
In 2002, a British programmer and inventor, Nick Pelling coined the term gamification in which the advent gaming concepts are introduced in real life to achieve better user interface design amongst electronic transactions. Up to the year 2011, the gamification is not popular until Gartner realized the advantages of gamification and introduced the concept into his Hype Cycle list.
Gartner, a leading IT research organization, has predicted that more than 50 percent of the organization will be replaced with gamification in the future and will be implemented in almost every part of the life cycle.
In recent years, design features captured through gamification has bought notable changes in the field of academia as well as industry. However, due to the dearth in comprehensive understanding and lack of resources has made gamification to fail.
This limitation is overcome through the latest gaming design strategies comprising simple mechanics such as leaderboards, points, and badges
>Read the complete article: Gamification Tied to Business Needs
Web Development has advanced over a period of time
Taking a Mobile First Strategy to Web-development
Changing trends in web design via web development
Java's Ongoing Popularity
The Era of Modern Web Development
I personally prefer Java as a first language. Put non-technically, it is a lot less convenient, so you get a more realistic idea of how computers work. Nonetheless, Python is an amazing language (with convenience as one of it’s values) so it makes sense as a gentle introduction. It’s also a popular language for data science and machine learning, so it’s great to have experience with.
The Python Language Interpreter: when you write some code in a text file and save it as a .py file, the Python interpreter is what turns that code into commands that your computer can then actually preform. This is necessary.
An Integrated Development Environment (IDE): An IDE is like a helpful text editor for programming. Some basic features include auto-complete, typo and mistake catching, and automatic text coloring to make some parts of your code easier to find. This is optional but highly recommended.
Some learning resources: We’re going to need something about programming basics, problem solving in computer science, using an API, learning how to use google and stackexchange, data types, control structures, and then maybe an object-oriented programming intro, and eventually all the neat advanced features of the python language. Then we need to learn how to use Numpy (for scientific computing), Pandas (for easy data storage), and Tensorflow (machine learning!). Add some handy cheat-sheets for python, numpy, pandas, and Tensorflow, and we’re good to go.
Other posts will adress download, installation, and resources.
Like I said up above, we need to know how to do the following. Save this and make it a checklist.
Learn to use google to answer questions about installing or using python, any packages, or computer science.
This also includes getting to know how to search Stackexchange, the website for coding questions n’ stuff.
How to install python 3 and get set up
How to install an IDE like Eclipse (with PyDev), IDLE, or Notebook++.
Programming basics: how does python work? What does the language look like? How does tabbing work?
Understand basic logic, including AND, OR, XOR, NOR, NAND, XNOR, Implies, and If…Else statements.
Variables: what are they, how do I set one and change it?
Basic math in Python.
Data types: what kinds of variables can I have? How does my computer store data? How do I use those types of data? What are the key commands and operations I know how to do?
Control structures: if, else, elif, for loops, while loops, break, continue
Methods! What are they, how do I make one, what can I do with it?
The open() command, the all() command, other neat built-in methods
<function name>= lambda <your variables>: <single line method>
Problem solving in computer science: now do fizbuzz.
What’s a package?
Importing packages, installing packages you don’t have with PIP
Using an API: how do I find one and how do I read it?
object-oriented programming in Python: what’s a class, how do I make one, how do I reference and instantiate one, methods, class vars, etc
Error handling: how to do exceptions
All the neat advanced features of the python language: iterators, generators, list comprehensions, enumerate, range, assert, with…as, etc.
Read through the Numpy API (for scientific computing), data types, matrices, stats, methods, etc. A short detour through scikit would be helpful.
Read through Matplotlib.pyplot API, plotting, plotting options, histograms, scatterplots, etc.
Pandas (for easy data storage), data frames, series, built-in operations on columns and rows, loading from a CSV, saving as a csv, apply, etc
Tensorflow (machine learning!) For basic stuff, shoot for knowing how to use the estimator package, which is discussed elsewhere on this blog. Also get to know the nitty gritty, including tensors, layers, tensorboard, etc.
Big data is the collection of diverse sets of information which are produced in large volumes and have to be processed at speed. The collected data are analyzed to understand current trends, patterns and make the critical prediction. The potential of big data can’t be bounded.
When large data sets are accumulated, they need to be analyzed efficiently and adequately. Big data cannot be handled by the human workforce.
https://goo.gl/yt8REu
“Dr. Grey Walter was a neurologist, robotics pioneer, and a bit of a mad scientist. Living in Bristol, England in 1949, without the aid of modern day computer processors, he built reactive, autonomous robots that could wander about and avoid obstacles. Each robot had two simulated neurons, sufficient for them to display complex behavior. Significantly, Walter’s tortoises represent the first real world demonstration of artificial life.“
When AR and VR are integrated into the existing range of applications, rather than working on a single product, these technologies revolutionize the entire industry.
VR technology creates a virtual environment where users can interact using their mobile devices such as VR glasses. AR technology presents the virtual objects in the natural environment. For instance, VR glasses can be used to display information regarding something in the user’s field of vision.
Listed below are seven industries which are likely to be affected first by these two technologies: https://goo.gl/YvgmMs
Suggested: Cyber Security Predictions to Safeguard your Businesses
Source: Cybersecurity APAC
Cyber security and authentication have been under attack in recent months as, seemingly every other day, a new report of hackers gaining access to private or sensitive information comes to light. Just recently, more than 500 million passwords were stolen when Yahoo revealed its security was compromised.
Securing systems has gone beyond simply coming up with a clever password that could prevent nefarious computer experts from hacking into your Facebook account. The more sophisticated the system, or the more critical, private information that system holds, the more advanced the identification system protecting it becomes.
Fingerprint scans and iris identification are just two types of authentication methods, once thought of as science fiction, that are in wide use by the most secure systems. But fingerprints can be stolen and iris scans can be replicated. Nothing has proven foolproof from being subject to computer hackers.
“The principal argument for behavioral, biometric authentication is that standard modes of authentication, like a password, authenticates you once before you access the service,” said Abdul Serwadda a cybersecurity expert and assistant professor in the Department of Computer Science at Texas Tech University.
“Now, once you’ve accessed the service, there is no other way for the system to still know it is you. The system is blind as to who is using the service. So the area of behavioral authentication looks at other user-identifying patterns that can keep the system aware of the person who is using it. Through such patterns, the system can keep track of some confidence metric about who might be using it and immediately prompt for reentry of the password whenever the confidence metric falls below a certain threshold.”
One of those patterns that is growing in popularity within the research community is the use of brain waves obtained from an electroencephalogram, or EEG. Several research groups around the country have recently showcased systems which use EEG to authenticate users with very high accuracy.
However, those brain waves can tell more about a person than just his or her identity. It could reveal medical, behavioral or emotional aspects of a person that, if brought to light, could be embarrassing or damaging to that person. And with EEG devices becoming much more affordable, accurate and portable and applications being designed that allows people to more readily read an EEG scan, the likelihood of that happening is dangerously high.
“The EEG has become a commodity application. For $100 you can buy an EEG device that fits on your head just like a pair of headphones,” Serwadda said. “Now there are apps on the market, brain-sensing apps where you can buy the gadget, download the app on your phone and begin to interact with the app using your brain signals. That led us to think; now we have these brain signals that were traditionally accessed only by doctors being handled by regular people. Now anyone who can write an app can get access to users’ brain signals and try to manipulate them to discover what is going on.”
That’s where Serwadda and graduate student Richard Matovu focused their attention: attempting to see if certain traits could be gleaned from a person’s brain waves. They presented their findings recently to the Institute of Electrical and Electronics Engineers (IEEE) International Conference on Biometrics.
Brain waves and cybersecurity
Serwadda said the technology is still evolving in terms of being able to use a person’s brain waves for authentication purposes. But it is a heavily researched field that has drawn the attention of several federal organizations. The National Science Foundation (NSF), funds a three-year project on which Serwadda and others from Syracuse University and the University of Alabama-Birmingham are exploring how several behavioral modalities, including EEG brain patterns, could be leveraged to augment traditional user authentication mechanisms.
“There are no installations yet, but a lot of research is going on to see if EEG patterns could be incorporated into standard behavioral authentication procedures,” Serwadda said.
Assuming a system uses EEG as the modality for user authentication, typically for such a system, all variables have been optimized to maximize authentication accuracy. A selection of such variables would include:
The features used to build user templates.
The signal frequency ranges from which features are extracted
The regions of the brain on which the electrodes are placed, among other variables.
Under this assumption of a finely tuned authentication system, Serwadda and his colleagues tackled the following questions:
If a malicious entity were to somehow access templates from this authentication-optimized system, would he or she be able to exploit these templates to infer non-authentication-centric information about the users with high accuracy?
In the event that such inferences are possible, which attributes of template design could reduce or increase the threat?
Turns out, they indeed found EEG authentication systems to give away non-authentication-centric information. Using an authentication system from UC-Berkeley and a variant of another from a team at Binghamton University and the University of Buffalo, Serwadda and Matovu tested their hypothesis, using alcoholism as the sensitive private information which an adversary might want to infer from EEG authentication templates.
In a study involving 25 formally diagnosed alcoholics and 25 non-alcoholic subjects, the lowest error rate obtained when identifying alcoholics was 25 percent, meaning a classification accuracy of approximately 75 percent.
When they tweaked the system and changed several variables, they found that the ability to detect alcoholic behavior could be tremendously reduced at the cost of slightly reducing the performance of the EEG authentication system.
Motivation for discovery
Serwadda’s motivation for proving brain waves could be used to reveal potentially harmful personal information wasn’t to improve the methods for obtaining that information. It’s to prevent it.
To illustrate, he gives an analogy using fingerprint identification at an airport. Fingerprint scans read ridges and valleys on the finger to determine a person’s unique identity, and that’s it.
In a hypothetical scenario where such systems could only function accurately if the user’s finger was pricked and some blood drawn from it, this would be problematic because the blood drawn by the prick could be used to infer things other than the user’s identity, such as whether a person suffers from certain diseases, such as diabetes.
Given the amount of extra information that EEG authentication systems are able glean about the user, current EEG systems could be likened to the hypothetical fingerprint reader that pricks the user’s finger. Serwadda wants to drive research that develops EEG authentication systems that perform the intended purpose while revealing minimal information about traits other than the user’s identity in authentication terms.
Currently, in the vast majority of studies on the EEG authentication problem, researchers primarily seek to outdo each other in terms of the system error rates. They work with the central objective of designing a system having error rates which are much lower than the state-of-the-art. Whenever a research group develops or publishes an EEG authentication system that attains the lowest error rates, such a system is immediately installed as the reference point.
A critical question that has not seen much attention up to this point is how certain design attributes of these systems, in other words the kinds of features used to formulate the user template, might relate to their potential to leak sensitive personal information. If, for example, a system with the lowest authentication error rates comes with the added baggage of leaking a significantly higher amount of private information, then such a system might, in practice, not be as useful as its low error rates suggest. Users would only accept, and get the full utility of the system, if the potential privacy breaches associated with the system are well understood and appropriate mitigations undertaken.
But, Serwadda said, while the EEG is still being studied, the next wave of invention is already beginning.
“In light of the privacy challenges seen with the EEG, it is noteworthy that the next wave of technology after the EEG is already being developed,” Serwadda said. “One of those technologies is functional near-infrared spectroscopy (fNIRS), which has a much higher signal-to-noise ratio than an EEG. It gives a more accurate picture of brain activity given its ability to focus on a particular region of the brain.”
The good news, for now, is fNIRS technology is still quite expensive; however there is every likelihood that the prices will drop over time, potentially leading to a civilian application to this technology. Thanks to the efforts of researchers like Serwadda, minimizing the leakage of sensitive personal information through these technologies is beginning to gain attention in the research community.
“The basic idea behind this research is to motivate a direction of research which selects design parameters in such a way that we not only care about recognizing users very accurately but also care about minimizing the amount of sensitive personal information it can read,” Serwadda said.
Your Daily dose of Latest Technology Updates, news, articles across various Industry Sectors
267 posts