Accepted to ICANN 2019

The following paper has been accepted to the 28th International Conference on Artificial Neural Networks (ICANN 2019) as an oral presentation. In this paper, we propose a method to control the architecture complexity by adding the penalty term in the dynamic optimization method of neural network structures [Shirakawa et al. 2018].

Shota Saito and Shinichi Shirakawa: Controlling Model Complexity in Probabilistic Model-Based Dynamic Optimization of Neural Network Structures, 28th International Conference on Artificial Neural Networks (ICANN 2019) (Accepted as oral presentation) [arXiv]

Accepted to ICML 2019

The following paper has been accepted to the International Conference on Machine Learning (ICML 2019). In this paper, we generalize and improve the dynamic optimization method of neural network structures proposed in [Shirakawa et al. 2018], and realize the fast, robust, and widely-applicable neural architecture search method. [Reference] Youhei Akimoto, Shinichi Shirakawa, Nozomu Yoshinari, Kento Uchida, Shota Saito, and Kouhei Nishida: Adaptive Stochastic Natural Gradient Method for One-Shot Neural Architecture Search, Proceedings of the 36th International Conference on Machine Learning (ICML), Vol. [Read More]

Accepted to IEEE Transaction on Evolutionary Computation

The following paper has been accepted to IEEE Transaction on Evolutionary Computation. This paper is an extension of our GECCO 2018 paper, which theoretically analyzes the information geometric optimization (IGO) algorithm with isotropic Gaussian distribution on Convex Quadratic Functions.

Kento Uchida, Shinichi Shirakawa, Youhei Akimoto, “Finite-Sample Analysis of Information Geometric Optimization with Isotropic Gaussian Distribution on Convex Quadratic Functions,” IEEE Transactions on Evolutionary Computation (2019) (Accepted) [DOI]

New members!

The menbers page has been updated. Now, our laboratory has 12 graduate and five undergraduate students.

Members page

Accepted to Evolutionary Computation Journal

The following paper has been accepted to Evolutionary Computation Journal. This paper is an extension of our GECCO 2017 paper, which proposed an architecture search method of CNN using genetic programming.

Masanori Suganuma, Masayuki Kobayashi, Shinichi Shirakawa, and Tomoharu Nagao, “Evolution of Deep Convolutional Neural Networks Using Cartesian Genetic Programming,” Evolutionary Computation, MIT Press (Accepted)

GECCO 2018

We are going to present the following papers at Genetic and Evolutionary Computation Conference (GECCO) 2018 @ Kyoto.

Kento Uchida, Youhei Akimoto, Shinichi Shirakawa, “Analysis of Information Geometric Optimization with Isotropic Gaussian Distribution Under Finite Samples” (accepted as a full paper).

Shota Saito, Shinichi Shirakawa, Youhei Akimoto, “Embedded Feature Selection Using Probabilistic Model-Based Optimization” (to be presented at student workshop).

New members!

The menbers page has been updated. Now, our laboratory has eight graduate and six undergraduate students.

Members page

Our paper has been accepted to AAAI 2018!

Our paper, “Dynamic Optimization of Neural Network Structures Using Probabilistic Modeling”, has been accepted to the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18). This paper proposes the framework to dynamically optimize the neural network structure during the training using gradient-based method.

Best Paper Award in GECCO 2017

We won the Best Paper Award of Evolutionary Machine Learning (EML) Track in GECCO 2017! The title of the award paper is “A Genetic Programming Approach to Designing Convolutional Neural Network Architectures”. This paper achieved to find high-performance architectures of the convolutional neural networks by the genetic programming with the reasonable computational resource. [Reference] Masanori Suganuma, Shinichi Shirakawa, and Tomoharu Nagao: A Genetic Programming Approach to Designing Convolutional Neural Network Architectures, Genetic and Evolutionary Computation Conference 2017 (GECCO 2017), pp. [Read More]

New members!

The menbers page has been updated. Now, our laboratory has two graduate and six undergraduate students.

Members page