Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletter Hub
Free Learning
Arrow right icon
timer SALE ENDS IN
0 Days
:
00 Hours
:
00 Minutes
:
00 Seconds

Tech News - Data

1209 Articles
article-image-article-13-back-on-track-france-and-germany-join-hands-to-save-the-eus-copyright-directive
Melisha Dsouza
06 Feb 2019
3 min read
Save for later

Article 13 back on Track- France and Germany join hands to save the EU's Copyright Directive

Melisha Dsouza
06 Feb 2019
3 min read
Last month, The EU Copyright directive had been put on hold since the European Council (with representatives from all the member states) couldn’t establish a level ground for Article 13. 11 member nations voted against the law causing the final “trilogue” meeting (at which the law was supposed to be finalized) to be called off. According to the member states, Article 13 is ‘insufficiently protective of users’ rights.’ While most of the state governments remained in favor of Article 13, there was a certain disagreement about the details of this law. France and Germany couldn’t agree on which internet platforms should install upload filters to censor their users’ posts. The disagreement has been resolved, and the process of enacting the law is back in motion. This time- making the law even worse, says the EEF. This is because, after a lot of back and forth, Germany and France have come to an agreement that will possibly affect tons of smaller sites as well as the larger ones, with hardly any protection to sites that host user-generated content. Julia Reeda, a German politician and Member of the European Parliament, uploaded the Franco-German deal [PDF], that was leaked today and which shows that Upload filters must be installed by everyone except those services which fit all three of the following “extremely narrow criteria”: Available to the public for less than 3 years Annual turnover below €10 million Fewer than 5 million unique monthly visitors BoingBoing.net summarises the above saying, every single online platform where the public can communicate and that has been in operation for three years or more must immediately buy filters. The size of the company does not matter. Once a platform makes €5,000,000 in a year, it will be obligated to implement "copyright filters as well. And finally, every site must demonstrate that it has taken 'best efforts' to license anything that their users might conceivably upload. This means that any time a rightsholder offers the site a license for content that their users might use, they are obliged to buy it from them, at whatever price they name. The next step for this draft is that the national negotiators for EU member states approve the deal, and then a final vote in the European Parliament. If the law is finalised, there would be an enormous investment of money needed. Copyright filters will cost hundreds of millions of euros to develop and maintain. Besides the monetary aspect, the law may also block legitimate speech that probably uses copyrighted works to get a point across and is incorrectly identified as containing copyrighted works. The petition opposing this law is now the largest petition in European history. You can head over to Techdirt for more insights on this news. Lawmakers introduce new Consumer privacy bill and Malicious Deep Fake Prohibition Act to support consumer privacy and battle deepfakes Facebook hires top EEF lawyer and Facebook critic as Whatsapp privacy policy manager Russia opens civil cases against Facebook and Twitter over local data laws  
Read more
  • 0
  • 0
  • 9382

article-image-windows-ml-microsofts-planned-built-ai-platform-developers-windows-10
Sugandha Lahoti
08 Mar 2018
2 min read
Save for later

Windows ML: Microsoft’s planned built-in AI platform for developers in Windows 10

Sugandha Lahoti
08 Mar 2018
2 min read
Microsoft unveils plans to introduce more artificial intelligence and machine learning capabilities inside Windows 10. The next major Windows 10 update will now include a new AI platform, Windows ML. The new platform will enable developers to build machine-learning models, trained in Azure, right into their apps using Visual Studio and run them on their PCs. Here are a few noteworthy features: Windows ML has an abstraction layer at its core that can automatically optimize an application’s ML model for the underlying hardware. It adapts itself to every machine. So for example, if your computer includes a graphics card that supports Microsoft’s DirectX framework, Windows ML can use the software’s performance boosting features to enhance response times. On a less sophisticated machine, it might simply run AI models on the CPU. Developers can also import existing learning models from different AI platforms and run them locally on PCs and devices running on Windows 10. Microsoft researchers point out 3 benefits of using the Windows ML AI platform: Low latency, real-time results. Windows can perform AI evaluation tasks using the local processing capabilities of the PC, enabling real-time analysis of large local data such as images and video. Reduced operational costs. Developers can build affordable, end-to-end AI solutions that combine training models in Azure with deployment to Windows devices for evaluation. Flexibility. Developers can choose to perform AI tasks on the device or in the cloud based on what their customers and scenarios need. Microsoft also plans to provide support for specialized chips to power AI software. As part of the effort, the company is collaborating with Intel Corp. to make Windows ML compatible with its Movidius vision processing units. Developers can get an early look at the AI platform on Windows with Visual Studio Preview 15.7.  For all others, the Windows ML API in standard desktops apps and Universal Windows Apps will be available across all editions of Windows 10 this year. To read about all release features, have a look at the official Windows blog.
Read more
  • 0
  • 0
  • 9367

article-image-tensorflow-1-7-0-rc0-arrives-close-heels-tensorflow-1-6-0
Sugandha Lahoti
15 Mar 2018
2 min read
Save for later

Tensorflow 1.7.0-rc0 arrives close on the heels of Tensorflow 1.6.0!

Sugandha Lahoti
15 Mar 2018
2 min read
It’s only been a few days since we witnessed the release of Tensorflow 1.6.0, and now the first release candidate of Tensorflow 1.7.0 is already here! There are quite a few major features and improvements in this new release candidate. However, no breaking changes are unveiled as such. With Tensorflow 1.7.0-rc0, TensorBoard Debugger Plugin, the graphical user interface (GUI) of TensorFlow Debugger (tfdbg), is now in alpha. Also, Eager mode is moving out of contrib. Other additional major features include: EGraph rewrites emulating fixed-point quantization compatible with TensorFlow Lite are now supported by new tf.contrib.quantize package. Easily customize gradient computation available with tf.custom_gradient. New tf.contrib.data.SqlDataset provides an experimental support for reading a sqlite database as a Dataset Distributed Mutex / CriticalSection added to tf.contrib.framework.CriticalSection. Better text processing with tf.regex_replace. Easy, efficient sequence input with tf.contrib.data.bucket_by_sequence_length Apart from these, there is a myriad of bug fixes and small changes. Some of these include: MaxPoolGradGrad support is added for Accelerated Linear Algebra (XLA). CSE pass from Tensorflow is now disabled. tf.py_func now reports the full stack trace if an exception occurs. TPUClusterResolver now integrated with GKE's integration for Cloud TPUs. A new library added for statistical testing of samplers. Helpers added to stream data from the GCE VM to a Cloud TPU. ClusterResolvers are integrated with TPUEstimator. Metropolis_hastings interface unified with HMC kernel. LIBXSMM convolutions moved to a separate --define flag so that they are disabled by default. MomentumOptimizer lambda fixed. tfp.layers boilerplate reduced via programmable docstrings. auc_with_confidence_intervals, a method for computing the AUC and confidence interval with linearithmic time complexity added. regression_head now accepts customized link function, to satisfy the usage that user can define their own link function if the array_ops.identity does not meet the requirement. initialized_value and initial_value behaviors fixed for ResourceVariables created from VariableDef protos. TensorSpec added to represent the specification of Tensors. Constant folding pass is now deterministic. To know about other bug-fixes and changes visit the Tensorflow 1.7.0-rc0 Github Repo.
Read more
  • 0
  • 0
  • 9361

article-image-verizon-hosted-ericsson-2018-oss-bss-user-group-with-a-quest-for-easy-theme
Amrata Joshi
20 Nov 2018
3 min read
Save for later

Verizon hosted Ericsson 2018 OSS/BSS User Group with a ‘Quest For Easy’ theme

Amrata Joshi
20 Nov 2018
3 min read
On 14th and 15th November, Verizon hosted the Ericsson’s 2018 OSS/BSS (Operations Support Systems/ Business Support Systems) user Group conference. The theme of the conference was ‘Quest for easy.’ The conference included presentations, demos, panels discussions and meetings with service providers from all over the world. This year the conference was held in New York, USA and was a two-day long event. The participants got some wonderful insights from the customers who shared their OSS/BSS experiences. The attendees also got a chance to go through around ten demos across Ericsson’s OSS/BSS portfolio. They got enlightened by the idea of what ‘Quest For Easy’ can mean for consumers and enterprises, operations and businesses. They got some information on how can ‘Quest For Easy’ create an impact on service providers and the products and services they offer. Highlights from the speakers Experts from Ericsson made the event more interesting. Emanuele Iannetti, Head of Solution Area BSS gave some useful and latest information to the audience about the Ericsson BSS portfolio. Marton Sabli, Head of Solution and Service Readiness for Solution Area OSS, Rick Mallon, Head of BSS Catalog and Order Management and Mats Karlsson, Head of Solution Area OSS also gave talks. Marton Sabli spoke about the orchestration of network slices and service assurance in a multi-vendor environment. Mats Karlsson gave his insight on 5G, right from enablers to monetization and Rick Mallon explained how Catalog and Order Care help in achieving simplicity and automation. Mohit Gupta, Director of Product Management at Ericsson, explained the audience the importance of insight-driven automation. Insights from the panel discussions The panel discussions were interesting as the customers got a chance to speak about the challenges, opportunities and their overall point of view. These panel discussions were divided into two tracks. The first one was based on revenue management while the other one covered service management, orchestration, and analytics. Arthur D. Little led one of the panels in which the customers including Verizon, Sprint and T-Mobile shared their views on the challenges and opportunities related to 5G and IoT. They also shared their experiences with digital transformation and monetization. Gary G Fujinami, Director of Performance Analytics Development / Operations at Verizon, threw some light on challenges of management and orchestration in an NFV (Network Functions Platform) enabled multi-operator world. Read more about this news on Ericsson. OpenStack Foundation to tackle open source infrastructure problems, will conduct conferences under the name ‘Open Infrastructure Summit’ Tableau 2019.1 beta announced at Tableau Conference 2018 “We call on the UN to invest in data-driven predictive methods for promoting peace”, Nature researchers on the eve of ViEWS conference
Read more
  • 0
  • 0
  • 9359

article-image-uber-to-restart-its-autonomous-vehicle-testing-nine-months-after-the-fatal-arizona-accident
Natasha Mathur
20 Dec 2018
3 min read
Save for later

Uber to restart its autonomous vehicle testing, nine months after the fatal Arizona accident

Natasha Mathur
20 Dec 2018
3 min read
It was back in March this year when a self-driving car by Uber killed a pedestrian, a 49-year-old Elaine Herzberg, in Tempe, Arizona. Uber, who had to halt the on-road testing of its autonomous vehicles after the incident, got the permission granted again to restart the testing yesterday. The authorization letter by the Pennsylvania Department of Transportation (PennDOT) confirmed that Uber will resume its on-road testing of self-driving cars in Pittsburgh. As per the details of the accident’s investigation, Rafaela Vasquez, the backup driver, had looked down at his phone 204 times during a course of a 43-minute test drive. After the accident, Uber had to halt all of its autonomous vehicle testing operations in Pittsburgh, Toronto, San Francisco, and Phoenix. Additionally, a shocking revelation was made last week by an Uber manager, Robbie Millie, who said that he tried to warn the company’s top executives about the danger, a few days before the fatal Arizona accident. According to Robbie Miller, a manager in the testing-operations group, he had sent an email to Uber’s top execs, where he warned them about the dangers related to the software powering Uber’s prototype “robo-taxis”. He also said that he warned them about the human backup drivers in the vehicles who hadn’t been properly trained to do their jobs efficiently. Other than that, Uber recently released its Uber safety report, where the company mentioned that it is committed to “anticipating and managing risks” that come with on-road testing of autonomous vehicles, however, it cannot guarantee to eliminate all of the risks involved. “We are deeply regretful for the crash in Tempe, Arizona, this March. In the hours following, we grounded our self-driving fleets in every city they were operating. In the months since, we have undertaken a top-to-bottom review of ATG’s safety approaches, system development, and culture. We have taken a measured, phased approach to return to on-road testing, starting first with manual driving in Pittsburgh”, said Uber. Although Uber has not released any details on when exactly it will be restarting its AV’s road testing, it says that it will only go back to on-road testing when it has implemented the improved processes. Moving on forward, Uber will make sure to always have two employees at the front seat of its self-driving cars at all times. There’s also going to be an automatic braking system enabled to strictly monitor the safety of the employees within these self-driving cars. Uber’s new family of AI algorithms sets records on Pitfall and solves the entire game of Montezuma’s Revenge Uber announces the 2019 Uber AI Residency Uber posted a billion dollar loss this quarter. Can Uber Eats revitalize the Uber growth story
Read more
  • 0
  • 0
  • 9357

article-image-google-opensources-tensorflow-gan-tfgan-library-for-generative-adversarial-networks-neural-network-model
Abhishek Jha
13 Dec 2017
11 min read
Save for later

Generative Adversarial Networks: Google open sources TensorFlow-GAN (TFGAN)

Abhishek Jha
13 Dec 2017
11 min read
If you have played the game Prince of Persia, you know what it is like defending yourself from the ‘shadow’ which tries to kill you. It’s a conundrum: If you kill the shadow you die; if you don’t do anything, you definitely die! For all its merits, Generative Adversarial Networks, or GAN, has faced a similar problem with differentiation. Most deep learning experts who endorse GAN mix their support with a little bit of caution – there is a stability issue! You may call it a holistic convergence problem. Both discriminator and generator are at loggerheads, while still being dependant on each other for efficient training. If one of them fails, the entire system fails. And you have got to ensure they don’t explode. The Prince of Persia is an interesting concept! To begin with, Neural Networks were designed to replicate human brain (albeit, artificially). They have succeeded in recognizing objects and processing natural languages. But to think and act like humans at that neurological level – let us admit it’s a far cry still. Which is why Generative Adversarial Networks became a hot topic in machine learning. It’s a relatively new architecture, but have gone on to revolutionize deep learning by accurately modeling real world data in ways better than any other model has done before. After all, they came up with a new model for training a neural net, with not one but two independent nets that work separately (and act as adversaries!) as Discriminator and Generator. Such a new architecture for an unsupervised neural network yields far better performance when compared to traditional nets. But the fact is, we have barely scratched the surface. Challenge is to train GAN here onwards. It comes with its own problems, such as failing to differentiate how many of a particular object should occur at a location, failing to adapt to 3D objects (it doesn’t understand the perspectives of frontview and backview), not being able to understand real-life holistic structures, etc. Substantial research has been taking place to take care of these problems. New models have been proposed to give more accurate results than previous techniques. Now  Google intends to make the Generative Adversarial Networks easier to experiment with! They have just open sourced TFGAN, a lightweight TensorFlow library designned to make it easy to train and evaluate GANs. [embed width="" height=""]https://www.youtube.com/watch?v=f2GF7TZpuGQ[/embed] According to Google, TFGAN provides the infrastructure to easily train a GAN, provides well-tested loss and evaluation metrics, and gives easy-to-use examples that highlight the expressiveness and flexibility of TFGAN. "We’ve also released a tutorial that includes a high-level API to quickly get a model trained on your data," Google said in its announcement. Source: research.googleblog.com The above image demonstrates the effect of an adversarial loss on image compression. The top row shows image patches from the ImageNet dataset. The middle row shows the results of compressing and uncompressing an image through an image compression neural network trained on a traditional loss. The bottom row shows the results from a network trained with a traditional loss and an adversarial loss. The GAN-loss images are sharper and more detailed, even if they are less like the original. TFGAN offers simple function calls for majority of GAN use-cases (where users can run a model in a few lines of code), but it's also built in a modular way that covers sophisticated GAN designs. "You can just use the modules you want — loss, evaluation, features, training, etc. are all independent. TFGAN’s lightweight design also means you can use it alongside other frameworks, or with native TensorFlow code," Google says, adding that GAN models written using TFGAN will easily benefit from future infrastructure improvements. That users can select from a large number of already-implemented losses and features without having to rewrite their own. Most importantly, Google is assuring us that the code is well-tested: "You don’t have to worry about numerical or statistical mistakes that are easily made with GAN libraries." Source: research.googleblog.com Most neural text-to-speech (TTS) systems produce over-smoothed spectrograms. When applied to the TacotronTTS system, Google says, a GAN can recreate some of the realistic-texture reducing artifacts in the resulting audio. And then, there is no harm in reiterating that when Google has open sourced a project, it must be absolute production ready! "When you use TFGAN, you’ll be using the same infrastructure that many Google researchers use, and you’ll have access to the cutting-edge improvements that we develop with the library," the tech giant added. To Start With import tensorflow as tf tfgan = tf.contrib.gan Why TFGAN? Easily train generator and discriminator networks with well-tested, flexible library calls. You can mix TFGAN, native TF, and other custom frameworks Use already implemented GAN losses and penalties (ex Wasserstein loss, gradient penalty, mutual information penalty, etc) Monitor and visualize GAN progress during training, and evaluate them Use already-implemented tricks to stabilize and improve training Develop based on examples of common GAN setups Use the TFGAN-backed GANEstimator to easily train a GAN model Improvements in TFGAN infrastructure will automatically benefit your TFGAN project Stay up-to-date with research as we add more algorithms What are the TFGAN components? TFGAN is composed of several parts which were designed to exist independently. These include the following main pieces (explained in detail below). core: provides the main infrastructure needed to train a GAN. Training occurs in four phases, and each phase can be completed by custom-code or by using a TFGAN library call. features: Many common GAN operations and normalization techniques are implemented for you to use, such as instance normalization and conditioning. losses: Easily experiment with already-implemented and well-tested losses and penalties, such as the Wasserstein loss, gradient penalty, mutual information penalty, etc evaluation: Use Inception Score or Frechet Distance with a pretrained Inception network to evaluate your unconditional generative model. You can also use your own pretrained classifier for more specific performance numbers, or use other methods for evaluating conditional generative models. examples and tutorial: See examples of how to use TFGAN to make GAN training easier, or use the more complicated examples to jumpstart your own project. These include unconditional and conditional GANs, InfoGANs, adversarial losses on existing networks, and image-to-image translation. Training a GAN model Training in TFGAN typically consists of the following steps: Specify the input to your networks. Set up your generator and discriminator using a GANModel. Specify your loss using a GANLoss. Create your train ops using a GANTrainOps. Run your train ops. There are various types of GAN setups. For instance, you can train a generator to sample unconditionally from a learned distribution, or you can condition on extra information such as a class label. TFGAN is compatible with many setups, and a few are demonstrated below: Examples Unconditional MNIST generation This example trains a generator to produce handwritten MNIST digits. The generator maps random draws from a multivariate normal distribution to MNIST digit images. See 'Generative Adversarial Networks' by Goodfellow et al. # Set up the input. images = mnist_data_provider.provide_data(FLAGS.batch_size) noise = tf.random_normal([FLAGS.batch_size, FLAGS.noise_dims]) # Build the generator and discriminator. gan_model = tfgan.gan_model( generator_fn=mnist.unconditional_generator, # you define discriminator_fn=mnist.unconditional_discriminator, # you define real_data=images, generator_inputs=noise) # Build the GAN loss. gan_loss = tfgan.gan_loss( gan_model, generator_loss_fn=tfgan_losses.wasserstein_generator_loss, discriminator_loss_fn=tfgan_losses.wasserstein_discriminator_loss) # Create the train ops, which calculate gradients and apply updates to weights. train_ops = tfgan.gan_train_ops( gan_model, gan_loss, generator_optimizer=tf.train.AdamOptimizer(gen_lr, 0.5), discriminator_optimizer=tf.train.AdamOptimizer(dis_lr, 0.5)) # Run the train ops in the alternating training scheme. tfgan.gan_train( train_ops, hooks=[tf.train.StopAtStepHook(num_steps=FLAGS.max_number_of_steps)], logdir=FLAGS.train_log_dir) Conditional MNIST generation This example trains a generator to generate MNIST images of a given class. The generator maps random draws from a multivariate normal distribution and a one-hot label of the desired digit class to an MNIST digit image. See 'Conditional Generative Adversarial Nets' by Mirza and Osindero. # Set up the input. images, one_hot_labels = mnist_data_provider.provide_data(FLAGS.batch_size) noise = tf.random_normal([FLAGS.batch_size, FLAGS.noise_dims]) # Build the generator and discriminator. gan_model = tfgan.gan_model( generator_fn=mnist.conditional_generator, # you define discriminator_fn=mnist.conditional_discriminator, # you define real_data=images, generator_inputs=(noise, one_hot_labels)) # The rest is the same as in the unconditional case. ... Adversarial loss This example combines an L1 pixel loss and an adversarial loss to learn to autoencode images. The bottleneck layer can be used to transmit compressed representations of the image. Neutral networks with pixel-wise loss only tend to produce blurry results, so the GAN can be used to make the reconstructions more plausible. See 'Full Resolution Image Compression with Recurrent Neural Networks' by Toderici et al for an example of neural networks used for image compression, and 'Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network' by Ledig et al for a more detailed description of how GANs can sharpen image output. # Set up the input pipeline. images = image_provider.provide_data(FLAGS.batch_size) # Build the generator and discriminator. gan_model = tfgan.gan_model( generator_fn=nets.autoencoder, # you define discriminator_fn=nets.discriminator, # you define real_data=images, generator_inputs=images) # Build the GAN loss and standard pixel loss. gan_loss = tfgan.gan_loss( gan_model, generator_loss_fn=tfgan_losses.wasserstein_generator_loss, discriminator_loss_fn=tfgan_losses.wasserstein_discriminator_loss, gradient_penalty=1.0) l1_pixel_loss = tf.norm(gan_model.real_data - gan_model.generated_data, ord=1) # Modify the loss tuple to include the pixel loss. gan_loss = tfgan.losses.combine_adversarial_loss( gan_loss, gan_model, l1_pixel_loss, weight_factor=FLAGS.weight_factor) # The rest is the same as in the unconditional case. ... Image-to-image translation This example maps images in one domain to images of the same size in a different dimension. For example, it can map segmentation masks to street images, or grayscale images to color. See 'Image-to-Image Translation with Conditional Adversarial Networks' by Isola et al for more details. # Set up the input pipeline. input_image, target_image = data_provider.provide_data(FLAGS.batch_size) # Build the generator and discriminator. gan_model = tfgan.gan_model( generator_fn=nets.generator, # you define discriminator_fn=nets.discriminator, # you define real_data=target_image, generator_inputs=input_image) # Build the GAN loss and standard pixel loss. gan_loss = tfgan.gan_loss( gan_model, generator_loss_fn=tfgan_losses.least_squares_generator_loss, discriminator_loss_fn=tfgan_losses.least_squares_discriminator_loss) l1_pixel_loss = tf.norm(gan_model.real_data - gan_model.generated_data, ord=1) # Modify the loss tuple to include the pixel loss. gan_loss = tfgan.losses.combine_adversarial_loss( gan_loss, gan_model, l1_pixel_loss, weight_factor=FLAGS.weight_factor) # The rest is the same as in the unconditional case. ... InfoGAN Train a generator to generate specific MNIST digit images, and control for digit style without using any labels. See 'InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets' for more details. # Set up the input pipeline. images = mnist_data_provider.provide_data(FLAGS.batch_size) # Build the generator and discriminator. gan_model = tfgan.infogan_model( generator_fn=mnist.infogan_generator, # you define discriminator_fn=mnist.infogran_discriminator, # you define real_data=images, unstructured_generator_inputs=unstructured_inputs, # you define structured_generator_inputs=structured_inputs) # you define # Build the GAN loss with mutual information penalty. gan_loss = tfgan.gan_loss( gan_model, generator_loss_fn=tfgan_losses.wasserstein_generator_loss, discriminator_loss_fn=tfgan_losses.wasserstein_discriminator_loss, gradient_penalty=1.0, mutual_information_penalty_weight=1.0) # The rest is the same as in the unconditional case. ... Custom model creation Train an unconditional GAN to generate MNIST digits, but manually construct the GANModel tuple for more fine-grained control. # Set up the input pipeline. images = mnist_data_provider.provide_data(FLAGS.batch_size) noise = tf.random_normal([FLAGS.batch_size, FLAGS.noise_dims]) # Manually build the generator and discriminator. with tf.variable_scope('Generator') as gen_scope: generated_images = generator_fn(noise) with tf.variable_scope('Discriminator') as dis_scope: discriminator_gen_outputs = discriminator_fn(generated_images) with variable_scope.variable_scope(dis_scope, reuse=True): discriminator_real_outputs = discriminator_fn(images) generator_variables = variables_lib.get_trainable_variables(gen_scope) discriminator_variables = variables_lib.get_trainable_variables(dis_scope) # Depending on what TFGAN features you use, you don't always need to supply # every `GANModel` field. At a minimum, you need to include the discriminator # outputs and variables if you want to use TFGAN to construct losses. gan_model = tfgan.GANModel( generator_inputs, generated_data, generator_variables, gen_scope, generator_fn, real_data, discriminator_real_outputs, discriminator_gen_outputs, discriminator_variables, dis_scope, discriminator_fn) # The rest is the same as the unconditional case. ... Google has allowed anyone to contribute to the github repositories to facilitate code-sharing among machine learning users. For more examples on TFGAN, see tensorflow/models on GitHub.
Read more
  • 0
  • 0
  • 9331
Unlock access to the largest independent learning library in Tech for FREE!
Get unlimited access to 7500+ expert-authored eBooks and video courses covering every tech area you can think of.
Renews at $19.99/month. Cancel anytime
article-image-blazingdb-announces-blazingsql-a-gpu-sql-engine-for-nvidias-open-source-rapids
Natasha Mathur
11 Oct 2018
2 min read
Save for later

BlazingDB announces BlazingSQL , a GPU SQL Engine for NVIDIA’s open source RAPIDS

Natasha Mathur
11 Oct 2018
2 min read
The BlazingDB team announced a new and free version of BlazingDB’s query execution engine for RAPIDS open-source software by NVIDIA, called BlazingSQL, yesterday. BlazingSQL provides query datasets from enterprise Data Lakes directly into GPU memory as a GPU DataFrame (GDF). GPU DataFrame (GDF) is a project that offers support for interoperability between GPU applications. It also defines a common GPU in-memory data layer. To provide this data lake integration, and to enable SQL queries on the software, critical open-source libraries were built inside the RAPIDS open-source software. These libraries were then layered on a series of modules from BlazingDB. GDF provides users with PyGDF or Dask_GDF that offers a simple interface similar to the Pandas DataFrame.               BlazingSQL BlazingSQL also allows Python developers to execute SQL queries directly on the flat files that exist inside the distributed file systems. Moreover, it comes with cuML and cuDNN that comprises GPU-accelerated machine learning and deep learning libraries using GDFs. The GPU DataFrame offers developers the ability to run complete machine learning workloads inside the GPU memory. This reduces the cost of data exchange between different tools, as well as the transfer overhead over the PCIe bus. The BlazingDB team has given a demo and binary roadmap for the upcoming BlazingSQL releases. BlazingSQL 0.1 uses PyBlazing connection to execute SQL queries on GDFs loaded by the PyGDF API. It will be releasing in the next couple of weeks before 25th October. BlazingSQL 0.2 involves the integration of BlazingDB’s FileSystem API. This adds the ability to directly query flat files inside the existing distributed file systems. This will be releasing sometime between 25th October to 8th November. BlazingSQL 0.3 comprises the integration of the distributed scheduler so SQL queries are fanned out across multiple GPUs and servers. This will be releasing between 8th November and 30th November.  Finally, the BlazingSQL 0.4 will have Integration of the distributed, multi-layered cache. The release date for BlazingSQL 0.4 hasn’t been assigned but it is expected to release in 2018. For more information, check out the official BlazingDB blog post. Introducing Watermelon DB: A new relational database to make your React and React Native apps highly scalable MariaDB acquires Clustrix to give database customers ‘freedom from Oracle lock-in’ RxDB 8.0.0, a reactive, offline-first, multiplatform database for JavaScript released!
Read more
  • 0
  • 0
  • 9329

article-image-canadian-court-rules-out-ubers-arbitration-process-calls-it-unconscionable-and-invalid
Sugandha Lahoti
04 Jan 2019
3 min read
Save for later

Canadian court rules out Uber’s arbitration process; calls it “unconscionable” and “invalid”

Sugandha Lahoti
04 Jan 2019
3 min read
The Canadian top court has slammed Uber’s arbitration process allowing Uber drivers to turn to Canadian courts for resolving their disputes with Uber. According to Uber’s previous policy, Uber drivers and employees had to resolve their complaints through an international mediation process in the Netherlands which costed drivers US$14,500. In a rule released on Wednesday, a panel of three judges with the Court of Appeal for Ontario concluded that this arbitration clause in Uber’s driver services agreement was “unconscionable” and “invalid”. “It can be safely concluded that Uber chose this arbitration clause in order to favour itself and thus take advantage of its drivers who are clearly vulnerable to the market strength of Uber,” the ruling said. Uber considers its drivers as contractual workers instead of employees and hence denies basic worker rights to them such as sick leaves and minimum wages. Drivers protested and proposed class-action lawsuit to declare drivers as employees, not independent contractors. They demanded minimum wage, overtime and vacation pay claiming $400 million in damages. Uber argued that this lawsuit can’t proceed in Canada due to the arbitration clause. A lower court agreed, but the panel of three appeal court judges reversed the decision. The court found this clause improper due to two reasons. First, it is an illegal contracting out of an employment standard under the Employment Standards Act. Second, the clause is immoral considering the inequality of bargaining power between Uber and its drivers. “This decision confirms that employment laws actually matter in Ontario, and that you cannot deprive workers of their legal rights under the Ontario Employment Standards Act by sending them 6,000 km overseas to enforce those rights at exorbitant personal cost,” told lawyer Lior Samfiru who represents the proposed class-action plaintiffs and a partner at Samfiru Tumarkin LLP to Financial Post. “I think the message here is for companies … if you’re going to operate in Ontario, if you’re going to operate in Canada, you have to abide by our laws,” Samfiru said. “You have to play by the same rules as everyone else.” Uber Canada has released a statement saying that it is currently reviewing the court’s decision and is “proud to offer a flexible earning opportunity to tens of thousands of drivers throughout Ontario.” This news first appeared on Financial Post. Uber to restart its autonomous vehicle testing, nine months after the fatal Arizona accident Uber manager warned the leadership team of the inadequacy of safety procedures in their prototype robo-taxis early March, reports The Information Uber fined by British ICO and Dutch DPA for nearly $1.2m over a data breach from 2016.
Read more
  • 0
  • 0
  • 9298

article-image-youtube-disables-all-comments-on-videos-featuring-children-in-an-attempt-to-curb-predatory-behavior-and-appease-advertisers
Sugandha Lahoti
01 Mar 2019
3 min read
Save for later

YouTube disables all comments on videos featuring children in an attempt to curb predatory behavior and appease advertisers

Sugandha Lahoti
01 Mar 2019
3 min read
YouTube has disabled all comments from its videos featuring young children in order to curb the spread of pedophiles who are using YouTube to trade clips of young girls in states of undress. This issue was first discovered, when Matt Watson, a video blogger, posted a 20-minute clip detailing how comments on YouTube were used to identify certain videos in which young girls were in activities that could be construed as sexually suggestive, such as posing in front of a mirror and doing gymnastics. Youtube’s content regulation practices have been in the spotlight in recent years. Last week, YouTube received major criticism for recommending videos of minors and allowing pedophiles to comment on these posts, with a specific time stamp of the video of when an exposed private part of the young child was visible. YouTube was also condemned for monetizing these videos allowing advertisements for major brands like Nestle, Fortnite, Disney, Fiat, Fortnite, L’Oreal, Maybelline, etc to be displayed on these videos. Following this news, a large number of companies have suspended their advertising spending from YouTube and refused to do so until YouTube took strong actions. In the same week, YouTube told Buzzfeed News that it is demonetizing channels that promote anti-vaccination content. YouTube said that this type of content does not align with its policy and called it “dangerous and harmful” content. Actions taken by YouTube YouTube said that it will now disable comments worldwide on almost all videos of minors by default. It said the change would take effect over several months. This will include videos featuring young and older minors that could be at risk of attracting predatory behavior. They are further introducing new comments classifier powered by machine learning that will identify and remove twice as many predatory comments as the old one. YouTube has also banned videos that encourage harmful and dangerous challenges. “We will continue to take actions on creators who cause egregious harm to the community”, they wrote in a blog post. "Nothing is more important to us than ensuring the safety of young people on the platform," said YouTube chief executive Susan Wojcicki on Twitter. https://twitter.com/SusanWojcicki/status/1101182716593135621 Despite her apologetic comments, she was on the receiving end of a brutal backlash with people asking her to resign from the organization. https://twitter.com/g8terbyte/status/1101221757233573899 https://twitter.com/KamenGamerRetro/status/1101186868052398080 https://twitter.com/SpencerKarter/status/1101305878014242822 The internet is slowly becoming a harmful place for young tweens. Not just Youtube, recently, TikTok, the popular video-sharing app which is a rage among tweens, was accused of illegally collecting personal information from children under 13. It was fined $5.7m by the US Federal Trade Commission. TikTok has now implemented features to accommodate younger US users in a limited, separate app experience that has additional safety and privacy protections. Similar steps have, however, not been implemented across their global operations. Nestle, Disney, Fortnite pull out their YouTube ads from paedophilic videos as YouTube’s content regulation woes continue. Youtube promises to reduce recommendations of ‘conspiracy theory’. Ex-googler explains why this is a ‘historic victory’. Is the YouTube algorithm’s promoting of #AlternativeFacts like Flat Earth having a real-world impact?
Read more
  • 0
  • 0
  • 9269

article-image-uber-manager-warned-the-leadership-team-of-the-inadequacy-of-safety-procedures-in-their-prototype-robo-taxis-early-march-reports-the-information
Sugandha Lahoti
13 Dec 2018
3 min read
Save for later

Uber manager warned the leadership team of the inadequacy of safety procedures in their prototype robo-taxis early March, reports The Information

Sugandha Lahoti
13 Dec 2018
3 min read
In a fatal accident on March 19, Uber’s prototype self-driving car struck and killed a pedestrian in Arizona. This incident raised alarms about safety problems in self-driving tech and Uber was criticized. In a shocking revelation made yesterday, The Information reported, that days before the fatal accident, an Uber manager tried to warn the company’s top executives about the danger. Robbie Miller, a manager in the testing-operations group, sent an email to Eric Meyhofer, the head of Uber’s autonomous vehicle unit, Jon Thomason, VP of software, and five other executives and lawyers on March 13. He warned them about the dangers of the software powering the company’s prototype robo-taxis. He also warned that the human backup drivers in the vehicles weren’t properly trained to do their jobs, the Information reports. What did Miller’s email say In his email, Miller pointed to an incident in November 2017, when an Uber car had nearly caused a crash. He prepared a report and urged the Uber team to investigate but was ignored. He was told that “incidents like that happen all of the time." Per Miller, “A car was damaged nearly every other day in February," Miller said. "We shouldn’t be hitting things every 15,000 miles." Miller was part of Uber's self-driving truck project, which he described as having relatively good safety procedures. The other projects focused on cars, and Miller argued that its safety procedures were extremely inadequate. In his report, Miller mentioned several ways to improve safety. He suggested Uber put two people in every vehicle. The driver should focus on the road while the other passenger can monitor the driving software and log misbehavior. Miller also argued that Uber should drastically scale back its testing program. "I suspect an 85% reduction in fleet size wouldn’t slow development," he wrote. Moreover, he wanted Uber to take strict actions against the fleet in case of a car crash. Everyone involved in the self-driving car project from developers to safety drivers should be given the authority to ground the fleet if they see a safety problem. He also wanted more personnel to have access to Uber's incident reporting database. People on the internet expressed their disdain over Uber’s safety neglect and sided with Miller. https://twitter.com/dhh/status/1072972633308688384 https://twitter.com/amir/status/1072508806935076864 https://twitter.com/sudo_lindenberg/status/1072669780899958789 Responding to the Information’s report, Uber said that “the entire team is focused on safely and responsibly returning to the road in self-driving mode,” The company intends to eventually resume on-the-road self-driving testing, but it will do so “only when these improvements have been implemented and we have received authorization from the Pennsylvania Department of Transportation.” This story first appeared on The Information. Introducing AWS DeepRacer, a self-driving race car, and Amazon’s autonomous racing league to help developers learn reinforcement learning in a fun way. Uber fined by British ICO and Dutch DPA for nearly $1.2m over a data breach from 2016. Uber’s new family of AI algorithms sets records on Pitfall and solves the entire game of Montezuma’s Revenge
Read more
  • 0
  • 0
  • 9243
article-image-top-executives-leave-facebook-after-pivot-to-privacy-announcement
Fatema Patrawala
15 Mar 2019
4 min read
Save for later

Two top executives leave Facebook soon after the pivot to privacy announcement

Fatema Patrawala
15 Mar 2019
4 min read
Facebook’s top executives, Chris Cox, Head of Products and Chris Daniels, Head of Whatsapp, have announced their exit from the company. It marks yet another highest-level departure at the tech giant amid a controversial shift to combine its various social media platforms. Cox’s unexpected departure, which he and Zuckerberg announced in separate Facebook posts on Thursday, comes months after Cox was promoted in a major reorganization. In May last year, Cox was put in charge of Facebook’s “family of apps,” including Instagram, Messenger, WhatsApp and Facebook itself — which together have over 2.7 billion users worldwide. These apps have been distinct until recently, when Zuckerberg announced plans to unify them with a new focus on privacy. “It is with great sadness I share with you that after thirteen years, I’ve decided to leave the company,” Cox wrote in his post. “Since I was twenty-three, I’ve poured myself into these walls. This place will forever be a part of me.” Another longtime executive, Chris Daniels also announced his exit on Thursday. Chris moved upward in the reorganization last May and took over WhatsApp after running Internet.org, the company’s philanthropic project to promote global Internet access. “At this point, we have made real progress on many issues and we have a clear plan for our apps, centered around making private messaging, stories and groups the foundation of the experience, including enabling encryption and interoperability across our services,” Zuckerberg wrote. “As we embark on this next major chapter, Chris has decided now is the time to step back from leading these teams.” In his blog post, Zuckerberg said Cox had told him several years ago that he planned to move but that Cox decided to hold off on leaving until the company made more progress combating misinformation and Russian interference — controversies that erupted in the wake of the 2016 election. Zuckerberg said he does not plan to replace Cox. The role of integrating the apps will go to another longtime Zuckerberg deputy and former Head of Growth, Javier Olivan, he said. Cox, who dropped out of a Stanford University graduate degree program to work with Zuckerberg when the company had just 15 engineers, was widely seen as one of the most popular and capable executives at the social network. Cox was a sounding board for Zuckerberg on product ideas. He launched Facebook’s flagship scrolling news feed nearly a decade ago and ran human resources before he was promoted to run the Facebook app in 2014. Cox is one of many senior executives to leave Facebook since the controversies erupted. The company’s Head of Policy Elliot Schrage, its General Counsel Colin Stretch, its Chief Security Officer Alex Stamos, along with the heads of WhatsApp and Instagram and its top communications executive, had all announced their exits from the company in the last few years. But the highest profile departures, including Cox, are people in Zuckerberg’s inner circle who have been at Facebook since the earliest days of the company and became fabulously wealthy after the social network’s $104 billion public offering. Cox hinted at the challenges of recent years in his goodbye post. “For over a decade, I’ve been sharing the same message that Mark and I have always believed: social media’s history is not yet written, and its effects are not neutral,” he wrote. “It is tied up in the richness and complexity of social life. As its builders we must endeavor to understand its impact — all the good, and all the bad — and take up the daily work of bending it towards the positive, and towards the good. This is our greatest responsibility.” Facebook tweet explains ‘server config change’ for 14-hour outage on all its platforms Facebook under criminal investigations for data sharing deals: NYT report Facebook family of apps hits 14 hours outage, longest in its history
Read more
  • 0
  • 0
  • 9218

article-image-foundationdb-6-0-15-releases-with-multi-region-support-and-seamless-failover-management
Natasha Mathur
20 Nov 2018
3 min read
Save for later

FoundationDB 6.0.15 releases with multi-region support and seamless failover management

Natasha Mathur
20 Nov 2018
3 min read
The FoundationDB team released version 6.0.15 of its distributed, NoSQL database, yesterday. FoundationDB 6.0.15 explores new features such as multi-region support, seamless failover management, along with performance changes, and bug fixes. FoundationDB is an open source, multi-model datastore by Apple that lets you store multiple data types in a single database. All data in FoundationDB is safely stored, distributed, and replicated in the Key-Value Store component. FoundationDB offers high performance on commodity hardware, helping you support very heavy loads at a low cost. Let’s have a look at what’s new in FoundationDB 6.0.15. New features FoundationDB 6.0.15 offers native multi-region support that dramatically increases your database's global availability. This also offers greater control over how failover scenarios are managed. Seamless failover is now possible in FoundationDB 6.9.15, allowing your cluster to survive the loss of an entire region without any service interruption. These features can be further deployed so that clients experience low-latency, single-region writes. Support has been added for asynchronous replication to a remote DC with processes in a single cluster. This improves the asynchronous replication provided by fdbdr as servers can fetch data from the remote DC in case all the other replicas have been lost in one DC. Additional support has been added for synchronous replication of the transaction log to a remote DC. This makes sure that the remote DC need not contain any storage servers. The TLS plugin has been statically linked into the client and server binaries. There is no longer a need for a separate library. The fileconfigure command has been added to fdbcli which configures a database from a JSON document. Performance changes The master recovery time for clusters with large amounts of data has been significantly reduced. Recovery time has been significantly reduced for cases where rollbacks are executed on the memory storage engine. Clients can now update their key location cache much more efficiently after the reboots of storage servers. Multiple resolver configurations have been tuned to carry out the job balancing work more efficiently between each resolver. Bug Fixes Clusters that been configured to use TLS would get stuck, leading to all their CPUs getting used for opening new connections. This issue has been fixed now. The issue of TLS certificate reloading causing the TLS connections to drop until the processes were restarted has been fixed. The issue of Watches registered on a lagging storage server taking a long time to trigger has been fixed. Other Changes The capitalization of trace event names and attributes has been normalized. Memory requirements of the transaction log have been increased by 400MB. The replication factor in status JSON has been stored under redundancy_mode instead of redundancy.factor. The metric data_version_lag is replaced by data_lag.versions and data_lag.seconds. Several additional metrics have been added for the number of watches and mutation count and are exposed through status. For more information on FoundationDB 6.0.15, check out the official release notes. MongoDB switches to Server Side Public License (SSPL) to prevent cloud providers from exploiting its open source code BlazingDB announces BlazingSQL , a GPU SQL Engine for NVIDIA’s open source RAPIDS MongoDB acquires mLab to transform the global cloud database market and scale MongoDB Atlas
Read more
  • 0
  • 0
  • 9172

article-image-leaked-memo-reveals-that-facebook-has-threatened-to-pull-investment-projects-from-canada-and-europe-if-their-data-demands-are-not-met
Amrata Joshi
04 Mar 2019
4 min read
Save for later

Leaked memo reveals that Facebook has threatened to pull investment projects from Canada and Europe if their data demands are not met

Amrata Joshi
04 Mar 2019
4 min read
Facebook has threatened to pull investment projects from Canada and Europe if the lobbying demands stated by Sheryl Sandberg, COO at Facebook were not met, The Guardian reports. Facebook was planning to build a data center in Canada to create jobs. The leaked memo, as seen by CW and the Guardian reveals that the deal was to be made only if Christian Paradis, Canada's then minister of industry, sends a letter of reassurance to Sandberg. According to her, the letter should reassure Facebook that the existence of the data center would not be used by the country to extend its legal jurisdiction over non-Canadian data held by Facebook. Sandberg told the officials from the European Union and Canada that if she did not receive any reassurances, then Facebook will consider other options for investment and growth. On the same day, Facebook received the letter from Canada guaranteeing the independence of non-Canadian data. The EU is yet to give such an assurance. Because of the company’s relationship with the Irish government, Facebook was hoping to influence the EU as well. These confidential documents apparently got leaked online. They were filed under seal as part of a lawsuit in California between Facebook and an app developer, Six4Three. These confidential documents show a global lobbying operation by Facebook that targets legislators around the world, including countries like the U.K., United States, Canada, India, and Brazil. In a statement to Business Insider, Facebook said, "Like the other documents that were cherry-picked and released in violation of a court order last year, these by design tell one side of a story and omit important context. As we've said, these selective leaks came from a lawsuit where Six4Three, the creators of an app known as Pikinis, hoped to force Facebook to share information on friends of the app's users. These documents have been sealed by a Californian court so we're not able to discuss them in detail." According to Computer Weekly, one of the original reporters of the news,  Marne Levine, then Facebook's vice-president of global public policy, wrote in one memo, "Sheryl took a firm approach and outlined that a decision on the datacentre was imminent. She emphasized that if we could not get comfort from the Canadian government on the jurisdiction issue we had other options.” Levine also described in the leaked messages as to how the Facebook staff distracted aides to Paradis so that other lobbyists could initiate a discussion with the ministers directly. This made Levine get the mobile numbers of the three government ministers. According to Levine, Sheryl Sandberg got comfortable around former UK chancellor George Osborne. The motive was to make him speak out against EU data laws, according to the leaked internal memo. This news is a real eye-opener in terms of how Facebook operates, which might also be used as an inspiration by other tech companies in countries where their data demands are not being met. This also seems to be a winning situation for Facebook as it is not only getting its demands fulfilled but also receiving enough support from the government's end in doing it. “In a lot of ways Facebook is more like a government than a traditional company,” Facebook CEO Mark Zuckerberg has said in an interview. Well, it seems Mark Zuckerberg is on his point this time. The involvement of government is a matter of concern for most of the users. One of the users commented on HackerNews, “Just for a little context, I think it's worth mentioning that this news comes to light when Canadians are thinking quite a bit about companies lobbying the gov't, as a bit of a scandal is brewing with the current liberal gov't[0].” Another user commented, “The Canadians agreed to not regulate other countries data. This seems pretty reasonable. Why should the Canadian government regulate how an American tech company handles German data? It makes a lot more sense for each country to have jurisdiction over data from (1) its own citizens, (2) residents on its soil or (3) data physically stored on its soil.” Facebook announces ‘Habitat’, a platform for embodied ArtificiaI Intelligence research Facebook open sources Magma, a software platform for deploying mobile networks The Verge spotlights the hidden cost of being a Facebook content moderator, a role Facebook outsources to 3rd parties to make the platform safe for users
Read more
  • 0
  • 0
  • 9158
article-image-trick-or-treat-new-facebook-community-actions-for-users-to-create-petitions-and-connect-with-public-officials
Amrata Joshi
22 Jan 2019
5 min read
Save for later

Trick or Treat - New Facebook Community Actions for users to create petitions and connect with public officials

Amrata Joshi
22 Jan 2019
5 min read
Yesterday, Facebook launched a petition feature called Community Actions, which enables community users to request changes from their local and national elected officials and government agencies, as reported by TechCrunch. This feature has been rolled out to users across the United States. Users can create a petition, tag public officials or organizations, and also get their friends to support their cause. Supporters can discuss the topic related to a specific petition with fellow supporters on the page, and also create events and fundraisers. Facebook will display the number of supporters behind a Community Action, but users will be able to see the names of those they are friends with or can view pages or public figures. This feature comes with a one-click Support option, which is quite visible on the news feed and it reduces the time required for signing up. This, in turn, helps the organizations and individuals to maximize the size of their community. In a statement to TechCrunch, a Facebook spokesperson said, “Building informed and civically engaged communities is at the core of Facebook’s mission. Every day, people come together on Facebook to advocate for causes they care about, including by contacting their elected officials, launching a fundraiser, or starting a group. Through these and other tools, we have seen people marshal support for and get results on issues that matter to them. Community Action is another way for people to advocate for changes in their communities and partner with elected officials and government agencies on solutions.” Lately, Facebook has been working towards a number of features designed to get people more involved in their communities. Features such as Town Hall, which gives access to local officials and Candidate feature that allows politicians to pitch on camera, are few of the steps in this direction. According to TechCrunch, there are some limits wherein users can’t tag President Donald Trump or Vice President Mike Pence. This might prevent users from expressing themselves and putting up petitions for or against them. Though Facebook will use a combination of user flagging, proactive algorithmic detection, and human enforcers, the new feature might get misused in some way or the other. This feature could be used in a way to pressurize or bully politicians and bureaucrats. A major issue with this feature is that users can’t stand against a Community Action. The discussion feed might not include the negative points as only the supporters can discuss on the thread. But this might also lead trolls to falsely back them and disturb the entire discussion thread. In a statement to TechCrunch, Facebook said, “Users will have to share a Community Action to their own feed with a message of disapproval, or launch their own in protest.” The Community Actions might be used to spread some fake awareness and bring petitions which are not for the well-being of the users. If the support count gets manipulated, it might cause trouble as a wrong petition would get support. For example, a few of the communities could falsely manipulate users by using Facebook groups or message threads so that it would look like there’s much more support for a misleading cause. Another example is if a politician causes a community for backing and further manipulating votes based on their false posts and comment threads. With Facebook’s WhatsApp now working towards preventing the spread of fake news by restricting the forwards to 5 individuals or groups, Facebook’s Community Actions feature might work against it. Users are giving mixed reactions to this news. Some of the users seem to be excited about this new feature. One of the comments on HackerNews reads, “That's interesting. I see that Facebook develops more feature to support different initiatives (FB groups, charity pages) and even petition pages.” Some users think that Facebook would gather users’ data based on political views. This would help the company in organizing various ad campaigns and generating revenue. One of the users commented on HackerNews, “Facebook really doesn't care too much what the petitions are about, but is mostly interested in gathering more data on its users' political beliefs so they can allow domestic and foreign campaign spending groups to better target advertisements meant to change or reinforce those beliefs (or suppress civic participation of those with such beliefs) and increase FB's total share of campaign-related ad spend.” Some of the users don’t trust Facebook anymore and they think that the new features won’t be secure. Another comment on HackerNews reads, “I'm going to have a really hard time taking any new development coming out of Facebook as genuine, honest or non-privacy invasive. I simply do not foresee my opinion of Facebook, Zuckerberg or anyone still working there changing radically in the near future.” Others are not interested in any sort of political engagement on social media as they think the views would be manipulated there. Users are requesting for better sign up or verification process which would help keep the fake accounts away. FTC officials plan to impose a fine of over $22.5 billion on Facebook for privacy violations, Washington Post reports Facebook takes down Russian news agency, Sputnik’s pages for engaging in “coordinated inauthentic behavior” Facebook open sources Spectrum 1.0.0, an image processing library for better mobile image production
Read more
  • 0
  • 0
  • 9155

article-image-googles-chief-diversity-officer-danielle-brown-resigns-to-join-hr-tech-firm-gusto
Natasha Mathur
15 Apr 2019
4 min read
Save for later

Google’s Chief Diversity Officer, Danielle Brown resigns to join HR tech firm Gusto

Natasha Mathur
15 Apr 2019
4 min read
It was just a little over a week ago when Google released its diversity annual report for the year 2019. And last thursday, its chief diversity officer, Danielle Brown, who co-wrote the report with Melonie Parker, announced that she is leaving Google to join Gusto, a leading Denver and San Francisco based HR-tech firm. “I’m joining the team at Gusto...that’s on a mission to create a world where work empowers a better life. I’ll be leading the People team at a company that is all about people”, writes Brown in a LinkedIn post. https://twitter.com/dmbrown1/status/1116361389201739776 Brown is being replaced by Melonie Parker, who earlier served as the Global director of diversity, equity, and inclusion at Google. Brown had joined Google as the Chief Diversity Officer back in June 2017 and earlier worked at a similar profile at Intel. “Danielle has dedicated her career to helping foster humanity at work. Most recently, she served as vice president, employee engagement and chief diversity and inclusion officer at Google, where she focused on ensuring their workplace and culture were respectful, safe, and inclusive — values we hold paramount at Gusto. Danielle will be an incredible addition to the Gusto team”, said Josh Reeves, co-founder, and CEO, Gusto.   https://twitter.com/GustoHQ/status/1116360860492894210 Gusto serves 6 million small businesses all over the U.S. and provides small businesses with a full-service people platform. The platform provides business owners with all the features they need to build their team.   Eileen Naughton, Google VP of People Operations, confirmed Brown’s departure and told TechCrunch that she’s “grateful to Danielle for her excellent work over the past two years to improve representation in Google’s workforce and ensure an inclusive culture for everyone. We wish her all the best in her new role at Gusto”. https://twitter.com/JeffDean/status/1116567286372913152 Liz Fong Jones, a former Google Engineer, who left Google earlier this year in February, tweeted in response to the news of Brown’s departure, saying that it’s not a good sign for Google. She mentioned that Brown wasn't “always popular with execs and employees” but was a  “straight shooter”. https://twitter.com/lizthegrey/status/1116361831742881793 https://twitter.com/lizthegrey/status/1116362110743863298 Jones at her departure cited Google’s lack of leadership in response to the demands made by employees during the Google walkout in November 2018. She had also published a post on Medium, stating, ‘grave concerns’ related to strategic decisions made at Google and the way it ‘misused its power’. Brown hasn’t specified a reason for her departure from Google but wrote on her Linkedin post that “What if, in addition to trying to solve for employee engagement and inclusion within the biggest tech companies in the world, we tried to solve those critical needs for every local storefront, every new startup just getting off the ground, or every doctor’s office across our communities?” Google is facing a lot of controversies over its employee treatment and work culture. Just last week,  over 900 Google workers signed a letter urging Google for fair rights for its contract workers, who make up nearly 54% of the workforce. Google in response rolled out mandatory benefits for its TVCs including health care, paid sick leaves, tuition reimbursement, and minimum wage among others. Brown hasn’t spoken out yet anything regarding her experience within Google and writes that she’s “thrilled to join Gusto and advance its mission. I look forward to a future where work empowers a better life for all small businesses and their teams” Audience reaction to the news is largely positive with people congratulating Brown on her new role at Gusto. https://twitter.com/Katrina_HRM/status/1116522924578525184 https://twitter.com/nataliaenvy/status/1116391017458966528 Ian Goodfellow quits Google and joins Apple as a director of machine learning Google employees filed petition to remove anti-trans, anti-LGBTQ and anti-immigrant Kay Coles James from the AI Council Is Google trying to ethics-wash its decisions with its new Advanced Tech External Advisory Council?
Read more
  • 0
  • 0
  • 9128
Modal Close icon
Modal Close icon